var/home/core/zuul-output/0000755000175000017500000000000015067742630014537 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015067756332015507 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005067372515067756322017731 0ustar rootrootOct 03 12:51:03 crc systemd[1]: Starting Kubernetes Kubelet... Oct 03 12:51:03 crc restorecon[4572]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:03 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 12:51:04 crc restorecon[4572]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 03 12:51:04 crc restorecon[4572]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 03 12:51:04 crc kubenswrapper[4578]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 03 12:51:04 crc kubenswrapper[4578]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 03 12:51:04 crc kubenswrapper[4578]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 03 12:51:04 crc kubenswrapper[4578]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 03 12:51:04 crc kubenswrapper[4578]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 03 12:51:04 crc kubenswrapper[4578]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.642838 4578 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651465 4578 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651526 4578 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651535 4578 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651543 4578 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651550 4578 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651558 4578 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651564 4578 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651575 4578 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651583 4578 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651590 4578 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651597 4578 feature_gate.go:330] unrecognized feature gate: Example Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651605 4578 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651613 4578 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651621 4578 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651689 4578 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651698 4578 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651706 4578 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651714 4578 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651724 4578 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651735 4578 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651743 4578 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651750 4578 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651757 4578 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651765 4578 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651771 4578 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651782 4578 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651790 4578 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651798 4578 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651807 4578 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651832 4578 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651840 4578 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651848 4578 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.651855 4578 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652702 4578 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652720 4578 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652728 4578 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652734 4578 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652740 4578 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652746 4578 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652752 4578 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652760 4578 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652765 4578 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652774 4578 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652782 4578 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652789 4578 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652797 4578 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652804 4578 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652810 4578 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652816 4578 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652822 4578 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652827 4578 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652833 4578 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652838 4578 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652844 4578 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652849 4578 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652854 4578 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652861 4578 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652866 4578 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652874 4578 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652880 4578 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652886 4578 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652892 4578 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652897 4578 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652902 4578 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652907 4578 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652913 4578 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652918 4578 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652923 4578 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652928 4578 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652933 4578 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.652938 4578 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.653927 4578 flags.go:64] FLAG: --address="0.0.0.0" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.653956 4578 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.653971 4578 flags.go:64] FLAG: --anonymous-auth="true" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.653982 4578 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.653996 4578 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654004 4578 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654016 4578 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654025 4578 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654032 4578 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654039 4578 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654046 4578 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654055 4578 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654062 4578 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654069 4578 flags.go:64] FLAG: --cgroup-root="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654075 4578 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654081 4578 flags.go:64] FLAG: --client-ca-file="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654087 4578 flags.go:64] FLAG: --cloud-config="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654093 4578 flags.go:64] FLAG: --cloud-provider="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654100 4578 flags.go:64] FLAG: --cluster-dns="[]" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654110 4578 flags.go:64] FLAG: --cluster-domain="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654117 4578 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654124 4578 flags.go:64] FLAG: --config-dir="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654130 4578 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654138 4578 flags.go:64] FLAG: --container-log-max-files="5" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654156 4578 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654162 4578 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654169 4578 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654175 4578 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654182 4578 flags.go:64] FLAG: --contention-profiling="false" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654189 4578 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654195 4578 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654202 4578 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654208 4578 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654217 4578 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654223 4578 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654230 4578 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654236 4578 flags.go:64] FLAG: --enable-load-reader="false" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654242 4578 flags.go:64] FLAG: --enable-server="true" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654248 4578 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654256 4578 flags.go:64] FLAG: --event-burst="100" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654263 4578 flags.go:64] FLAG: --event-qps="50" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654269 4578 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654276 4578 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654283 4578 flags.go:64] FLAG: --eviction-hard="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654291 4578 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654298 4578 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654305 4578 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654314 4578 flags.go:64] FLAG: --eviction-soft="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654320 4578 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654327 4578 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654333 4578 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654340 4578 flags.go:64] FLAG: --experimental-mounter-path="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654347 4578 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654353 4578 flags.go:64] FLAG: --fail-swap-on="true" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654360 4578 flags.go:64] FLAG: --feature-gates="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654367 4578 flags.go:64] FLAG: --file-check-frequency="20s" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654374 4578 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654381 4578 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654388 4578 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654396 4578 flags.go:64] FLAG: --healthz-port="10248" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654404 4578 flags.go:64] FLAG: --help="false" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654412 4578 flags.go:64] FLAG: --hostname-override="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654420 4578 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654427 4578 flags.go:64] FLAG: --http-check-frequency="20s" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654433 4578 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654440 4578 flags.go:64] FLAG: --image-credential-provider-config="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654448 4578 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654456 4578 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654463 4578 flags.go:64] FLAG: --image-service-endpoint="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654470 4578 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654478 4578 flags.go:64] FLAG: --kube-api-burst="100" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654486 4578 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654496 4578 flags.go:64] FLAG: --kube-api-qps="50" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654504 4578 flags.go:64] FLAG: --kube-reserved="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654512 4578 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654519 4578 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654527 4578 flags.go:64] FLAG: --kubelet-cgroups="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654535 4578 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654543 4578 flags.go:64] FLAG: --lock-file="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654551 4578 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654560 4578 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654569 4578 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654583 4578 flags.go:64] FLAG: --log-json-split-stream="false" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654599 4578 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654609 4578 flags.go:64] FLAG: --log-text-split-stream="false" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654618 4578 flags.go:64] FLAG: --logging-format="text" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654661 4578 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654673 4578 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654681 4578 flags.go:64] FLAG: --manifest-url="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654689 4578 flags.go:64] FLAG: --manifest-url-header="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654701 4578 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654709 4578 flags.go:64] FLAG: --max-open-files="1000000" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654720 4578 flags.go:64] FLAG: --max-pods="110" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654729 4578 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654737 4578 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654745 4578 flags.go:64] FLAG: --memory-manager-policy="None" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654754 4578 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654762 4578 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654773 4578 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654781 4578 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654802 4578 flags.go:64] FLAG: --node-status-max-images="50" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654810 4578 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654819 4578 flags.go:64] FLAG: --oom-score-adj="-999" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654828 4578 flags.go:64] FLAG: --pod-cidr="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654836 4578 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654850 4578 flags.go:64] FLAG: --pod-manifest-path="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654876 4578 flags.go:64] FLAG: --pod-max-pids="-1" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654885 4578 flags.go:64] FLAG: --pods-per-core="0" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654894 4578 flags.go:64] FLAG: --port="10250" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654902 4578 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654911 4578 flags.go:64] FLAG: --provider-id="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654919 4578 flags.go:64] FLAG: --qos-reserved="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654928 4578 flags.go:64] FLAG: --read-only-port="10255" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654937 4578 flags.go:64] FLAG: --register-node="true" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654945 4578 flags.go:64] FLAG: --register-schedulable="true" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654953 4578 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654969 4578 flags.go:64] FLAG: --registry-burst="10" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654980 4578 flags.go:64] FLAG: --registry-qps="5" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.654989 4578 flags.go:64] FLAG: --reserved-cpus="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655001 4578 flags.go:64] FLAG: --reserved-memory="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655012 4578 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655022 4578 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655031 4578 flags.go:64] FLAG: --rotate-certificates="false" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655040 4578 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655048 4578 flags.go:64] FLAG: --runonce="false" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655057 4578 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655066 4578 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655075 4578 flags.go:64] FLAG: --seccomp-default="false" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655084 4578 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655092 4578 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655101 4578 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655110 4578 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655119 4578 flags.go:64] FLAG: --storage-driver-password="root" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655127 4578 flags.go:64] FLAG: --storage-driver-secure="false" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655135 4578 flags.go:64] FLAG: --storage-driver-table="stats" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655143 4578 flags.go:64] FLAG: --storage-driver-user="root" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655150 4578 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655159 4578 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655167 4578 flags.go:64] FLAG: --system-cgroups="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655176 4578 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655192 4578 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655201 4578 flags.go:64] FLAG: --tls-cert-file="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655208 4578 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655220 4578 flags.go:64] FLAG: --tls-min-version="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655230 4578 flags.go:64] FLAG: --tls-private-key-file="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655238 4578 flags.go:64] FLAG: --topology-manager-policy="none" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655245 4578 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655251 4578 flags.go:64] FLAG: --topology-manager-scope="container" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655258 4578 flags.go:64] FLAG: --v="2" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655269 4578 flags.go:64] FLAG: --version="false" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655278 4578 flags.go:64] FLAG: --vmodule="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655287 4578 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.655294 4578 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655517 4578 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655531 4578 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655541 4578 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655548 4578 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655555 4578 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655564 4578 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655572 4578 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655580 4578 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655594 4578 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655601 4578 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655610 4578 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655619 4578 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655651 4578 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655660 4578 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655667 4578 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655674 4578 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655680 4578 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655687 4578 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655694 4578 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655700 4578 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655707 4578 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655714 4578 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655721 4578 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655727 4578 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655734 4578 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655740 4578 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655747 4578 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655753 4578 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655762 4578 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655768 4578 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655775 4578 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655781 4578 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655788 4578 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655795 4578 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655801 4578 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655808 4578 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655814 4578 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655821 4578 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655831 4578 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655837 4578 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655849 4578 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655858 4578 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655866 4578 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655874 4578 feature_gate.go:330] unrecognized feature gate: Example Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655881 4578 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655888 4578 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655895 4578 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655915 4578 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655922 4578 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655930 4578 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655936 4578 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655944 4578 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655950 4578 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655957 4578 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655964 4578 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655972 4578 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655978 4578 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655985 4578 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655992 4578 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.655999 4578 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.656009 4578 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.656017 4578 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.656025 4578 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.656032 4578 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.656039 4578 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.656047 4578 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.656055 4578 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.656062 4578 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.656069 4578 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.656076 4578 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.656083 4578 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.656111 4578 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.669272 4578 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.669344 4578 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669516 4578 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669545 4578 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669560 4578 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669570 4578 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669582 4578 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669592 4578 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669603 4578 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669615 4578 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669625 4578 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669668 4578 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669680 4578 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669691 4578 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669701 4578 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669711 4578 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669721 4578 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669732 4578 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669742 4578 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669752 4578 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669764 4578 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669773 4578 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669783 4578 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669793 4578 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669807 4578 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669826 4578 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669839 4578 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669850 4578 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669861 4578 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669873 4578 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669884 4578 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669895 4578 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669907 4578 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669917 4578 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669927 4578 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669938 4578 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669952 4578 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669962 4578 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669972 4578 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669982 4578 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.669993 4578 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670003 4578 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670013 4578 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670027 4578 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670039 4578 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670049 4578 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670060 4578 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670071 4578 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670080 4578 feature_gate.go:330] unrecognized feature gate: Example Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670091 4578 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670101 4578 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670112 4578 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670125 4578 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670139 4578 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670151 4578 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670162 4578 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670175 4578 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670186 4578 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670196 4578 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670207 4578 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670217 4578 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670227 4578 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670237 4578 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670247 4578 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670257 4578 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670267 4578 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670279 4578 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670289 4578 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670303 4578 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670315 4578 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670327 4578 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670338 4578 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670351 4578 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.670369 4578 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670716 4578 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670739 4578 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670751 4578 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670762 4578 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670773 4578 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670783 4578 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670793 4578 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670803 4578 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670814 4578 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670825 4578 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670839 4578 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670852 4578 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670863 4578 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670875 4578 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670888 4578 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670900 4578 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670911 4578 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670921 4578 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670932 4578 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670942 4578 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670952 4578 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670962 4578 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670976 4578 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.670989 4578 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671001 4578 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671013 4578 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671025 4578 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671037 4578 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671047 4578 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671058 4578 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671069 4578 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671080 4578 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671090 4578 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671099 4578 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671121 4578 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671131 4578 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671141 4578 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671150 4578 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671160 4578 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671173 4578 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671185 4578 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671198 4578 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671211 4578 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671222 4578 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671232 4578 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671242 4578 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671253 4578 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671266 4578 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671276 4578 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671288 4578 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671300 4578 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671311 4578 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671346 4578 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671357 4578 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671367 4578 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671378 4578 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671387 4578 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671398 4578 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671407 4578 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671417 4578 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671428 4578 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671438 4578 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671447 4578 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671458 4578 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671468 4578 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671478 4578 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671493 4578 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671504 4578 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671514 4578 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671525 4578 feature_gate.go:330] unrecognized feature gate: Example Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.671542 4578 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.671561 4578 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.672002 4578 server.go:940] "Client rotation is on, will bootstrap in background" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.680530 4578 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.680735 4578 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.682814 4578 server.go:997] "Starting client certificate rotation" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.682862 4578 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.683073 4578 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-21 20:44:40.085209139 +0000 UTC Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.683194 4578 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1903h53m35.4020171s for next certificate rotation Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.714387 4578 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.717717 4578 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.738393 4578 log.go:25] "Validated CRI v1 runtime API" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.781436 4578 log.go:25] "Validated CRI v1 image API" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.784304 4578 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.789574 4578 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-03-12-44-59-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.789666 4578 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:42 fsType:tmpfs blockSize:0}] Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.809747 4578 manager.go:217] Machine: {Timestamp:2025-10-03 12:51:04.806588783 +0000 UTC m=+0.605061057 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2799886 MemoryCapacity:25199480832 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8 BootID:37bf46c8-7e72-46ed-94e0-9573cda7bef8 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599742464 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:42 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:56:e1:17 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:56:e1:17 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:9c:6e:e5 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:d0:98:04 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:bb:e8:e2 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:aa:91:be Speed:-1 Mtu:1496} {Name:eth10 MacAddress:f2:12:e8:b5:93:1c Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:9e:1f:d2:4c:f1:e9 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199480832 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.810058 4578 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.810339 4578 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.812496 4578 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.812804 4578 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.812863 4578 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.813145 4578 topology_manager.go:138] "Creating topology manager with none policy" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.813160 4578 container_manager_linux.go:303] "Creating device plugin manager" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.814362 4578 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.814411 4578 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.814803 4578 state_mem.go:36] "Initialized new in-memory state store" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.814924 4578 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.818609 4578 kubelet.go:418] "Attempting to sync node with API server" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.818668 4578 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.818706 4578 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.818725 4578 kubelet.go:324] "Adding apiserver pod source" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.818744 4578 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.824822 4578 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.824822 4578 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:04 crc kubenswrapper[4578]: E1003 12:51:04.824995 4578 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.165:6443: connect: connection refused" logger="UnhandledError" Oct 03 12:51:04 crc kubenswrapper[4578]: E1003 12:51:04.825005 4578 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.165:6443: connect: connection refused" logger="UnhandledError" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.825120 4578 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.827171 4578 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.829220 4578 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.830926 4578 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.830960 4578 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.830969 4578 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.830979 4578 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.830994 4578 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.831032 4578 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.831042 4578 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.831060 4578 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.831071 4578 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.831081 4578 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.831096 4578 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.831105 4578 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.832667 4578 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.833417 4578 server.go:1280] "Started kubelet" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.835666 4578 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.835871 4578 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.835984 4578 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 03 12:51:04 crc systemd[1]: Started Kubernetes Kubelet. Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.844009 4578 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.844054 4578 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.844098 4578 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.846164 4578 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.846194 4578 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.846541 4578 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 04:32:37.553767392 +0000 UTC Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.846708 4578 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1287h41m32.707069154s for next certificate rotation Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.846792 4578 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.847136 4578 server.go:460] "Adding debug handlers to kubelet server" Oct 03 12:51:04 crc kubenswrapper[4578]: E1003 12:51:04.851257 4578 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.851739 4578 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.851783 4578 factory.go:55] Registering systemd factory Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.851804 4578 factory.go:221] Registration of the systemd container factory successfully Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.851748 4578 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:04 crc kubenswrapper[4578]: E1003 12:51:04.851853 4578 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.165:6443: connect: connection refused" logger="UnhandledError" Oct 03 12:51:04 crc kubenswrapper[4578]: E1003 12:51:04.852248 4578 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.165:6443: connect: connection refused" interval="200ms" Oct 03 12:51:04 crc kubenswrapper[4578]: E1003 12:51:04.851908 4578 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.165:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186afc2ab7298973 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-03 12:51:04.833378675 +0000 UTC m=+0.631850899,LastTimestamp:2025-10-03 12:51:04.833378675 +0000 UTC m=+0.631850899,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.855739 4578 factory.go:153] Registering CRI-O factory Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.855862 4578 factory.go:221] Registration of the crio container factory successfully Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.855957 4578 factory.go:103] Registering Raw factory Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.856034 4578 manager.go:1196] Started watching for new ooms in manager Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.856775 4578 manager.go:319] Starting recovery of all containers Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.869715 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.869788 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.869996 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870012 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870024 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870036 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870047 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870060 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870074 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870093 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870106 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870119 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870128 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870141 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870150 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870160 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870171 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870182 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870191 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870208 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870218 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870229 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870238 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870264 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.870273 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871213 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871236 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871249 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871257 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871265 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871275 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871285 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871295 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871304 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871312 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871321 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871331 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871378 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871388 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871396 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871405 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871417 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871431 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871442 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871453 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871463 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871473 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871483 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871493 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871501 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871510 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871519 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871532 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871542 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871551 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871560 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871569 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871577 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871588 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.871596 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872010 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872026 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872035 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872045 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872057 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872065 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872074 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872082 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872091 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872100 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872109 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872118 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872128 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872136 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872145 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872155 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872164 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872173 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872182 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872193 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872203 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872211 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872219 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872229 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872238 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872247 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872255 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872264 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872274 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872286 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872298 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872310 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872322 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872337 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872350 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872363 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872376 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.872385 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873191 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873211 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873220 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873231 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873240 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873247 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873260 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873269 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873278 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873287 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873297 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873306 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873315 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873323 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873332 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873341 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873352 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873362 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873371 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873381 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873390 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873398 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873408 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873419 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873432 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873444 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873841 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873853 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873866 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873879 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873894 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873907 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873920 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.873933 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876250 4578 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876328 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876350 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876364 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876377 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876389 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876400 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876412 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876424 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876436 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876447 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876457 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876468 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876478 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876502 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876515 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876527 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876543 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876558 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876574 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876587 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876599 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876613 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876658 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876674 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876688 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876701 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876716 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876729 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876744 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876760 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876778 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876794 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876808 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876822 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876836 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876850 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876864 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876877 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876891 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876903 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876917 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876929 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876941 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876952 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876963 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876973 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876985 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.876995 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877005 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877015 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877026 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877038 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877048 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877057 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877069 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877083 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877097 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877110 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877122 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877136 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877153 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877166 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877179 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877190 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877200 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877209 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877219 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877228 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877237 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877246 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877261 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877272 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877282 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877291 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877301 4578 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877312 4578 reconstruct.go:97] "Volume reconstruction finished" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.877321 4578 reconciler.go:26] "Reconciler: start to sync state" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.879956 4578 manager.go:324] Recovery completed Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.901386 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.904149 4578 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.904216 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.904422 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.904439 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.905981 4578 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.906009 4578 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.906034 4578 state_mem.go:36] "Initialized new in-memory state store" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.907746 4578 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.907802 4578 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.907838 4578 kubelet.go:2335] "Starting kubelet main sync loop" Oct 03 12:51:04 crc kubenswrapper[4578]: E1003 12:51:04.907901 4578 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 03 12:51:04 crc kubenswrapper[4578]: W1003 12:51:04.908714 4578 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:04 crc kubenswrapper[4578]: E1003 12:51:04.908840 4578 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.165:6443: connect: connection refused" logger="UnhandledError" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.923619 4578 policy_none.go:49] "None policy: Start" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.924418 4578 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.924449 4578 state_mem.go:35] "Initializing new in-memory state store" Oct 03 12:51:04 crc kubenswrapper[4578]: E1003 12:51:04.952431 4578 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.987234 4578 manager.go:334] "Starting Device Plugin manager" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.987565 4578 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.987584 4578 server.go:79] "Starting device plugin registration server" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.988131 4578 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.988151 4578 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.988500 4578 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.988601 4578 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 03 12:51:04 crc kubenswrapper[4578]: I1003 12:51:04.988616 4578 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 03 12:51:04 crc kubenswrapper[4578]: E1003 12:51:04.999099 4578 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.008624 4578 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.008743 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.010205 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.010245 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.010257 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.010445 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.010821 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.010895 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.011790 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.011834 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.011849 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.012011 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.012219 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.012258 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.013041 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.013071 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.013087 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.013787 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.013841 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.013852 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.014028 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.014222 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.014253 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.015197 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.015218 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.015229 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.015321 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.015362 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.015394 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.015411 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.015585 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.015608 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.015622 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.015740 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.015768 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.017500 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.017520 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.017521 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.017529 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.017543 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.017555 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.017741 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.017767 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.018789 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.018852 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.018864 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:05 crc kubenswrapper[4578]: E1003 12:51:05.053436 4578 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.165:6443: connect: connection refused" interval="400ms" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.080277 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.080330 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.080362 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.080385 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.080407 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.080433 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.080454 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.080475 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.080498 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.080518 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.080541 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.080560 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.080580 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.080605 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.080654 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.088839 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.090560 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.090614 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.090625 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.090691 4578 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 12:51:05 crc kubenswrapper[4578]: E1003 12:51:05.091341 4578 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.165:6443: connect: connection refused" node="crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.182085 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.182202 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.182239 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.182277 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.182345 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.182346 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.182378 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.182752 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.182795 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.182831 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.182865 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.182897 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.182925 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.182954 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.182982 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.183010 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.182541 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.183341 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.182498 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.183410 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.182472 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.182433 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.183436 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.183452 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.183486 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.182505 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.183485 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.183511 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.183506 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.183511 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.291765 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.293338 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.293391 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.293404 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.293436 4578 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 12:51:05 crc kubenswrapper[4578]: E1003 12:51:05.294175 4578 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.165:6443: connect: connection refused" node="crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.356352 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.381943 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.389477 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.410139 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: W1003 12:51:05.410319 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-9731da0902b6256138ec7838b371ebb76e13b07484c4592648c0effea0a7330b WatchSource:0}: Error finding container 9731da0902b6256138ec7838b371ebb76e13b07484c4592648c0effea0a7330b: Status 404 returned error can't find the container with id 9731da0902b6256138ec7838b371ebb76e13b07484c4592648c0effea0a7330b Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.420862 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 03 12:51:05 crc kubenswrapper[4578]: W1003 12:51:05.439154 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-ec01d2592f8282c22d60740ba477b4293efba52bca476820223febca8ed62bc5 WatchSource:0}: Error finding container ec01d2592f8282c22d60740ba477b4293efba52bca476820223febca8ed62bc5: Status 404 returned error can't find the container with id ec01d2592f8282c22d60740ba477b4293efba52bca476820223febca8ed62bc5 Oct 03 12:51:05 crc kubenswrapper[4578]: W1003 12:51:05.441609 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-6b4e72775ca284b64c2113122d83995ccfb6de4cbbd985b0bd16b66a40cb6c02 WatchSource:0}: Error finding container 6b4e72775ca284b64c2113122d83995ccfb6de4cbbd985b0bd16b66a40cb6c02: Status 404 returned error can't find the container with id 6b4e72775ca284b64c2113122d83995ccfb6de4cbbd985b0bd16b66a40cb6c02 Oct 03 12:51:05 crc kubenswrapper[4578]: W1003 12:51:05.449319 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-34ae8f417e92fbd24bdf81fe407c67c4069ecbf7aad2db6d0d6a49c236dacc63 WatchSource:0}: Error finding container 34ae8f417e92fbd24bdf81fe407c67c4069ecbf7aad2db6d0d6a49c236dacc63: Status 404 returned error can't find the container with id 34ae8f417e92fbd24bdf81fe407c67c4069ecbf7aad2db6d0d6a49c236dacc63 Oct 03 12:51:05 crc kubenswrapper[4578]: W1003 12:51:05.450695 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-72cc7e49a39e23807e4c756837afa42f83a7e7df692dc86d9cba4bd86acaf6d9 WatchSource:0}: Error finding container 72cc7e49a39e23807e4c756837afa42f83a7e7df692dc86d9cba4bd86acaf6d9: Status 404 returned error can't find the container with id 72cc7e49a39e23807e4c756837afa42f83a7e7df692dc86d9cba4bd86acaf6d9 Oct 03 12:51:05 crc kubenswrapper[4578]: E1003 12:51:05.454567 4578 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.165:6443: connect: connection refused" interval="800ms" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.695282 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.697733 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.697764 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.697776 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.697800 4578 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 12:51:05 crc kubenswrapper[4578]: E1003 12:51:05.698158 4578 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.165:6443: connect: connection refused" node="crc" Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.837561 4578 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.911117 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"72cc7e49a39e23807e4c756837afa42f83a7e7df692dc86d9cba4bd86acaf6d9"} Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.912168 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6b4e72775ca284b64c2113122d83995ccfb6de4cbbd985b0bd16b66a40cb6c02"} Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.913318 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ec01d2592f8282c22d60740ba477b4293efba52bca476820223febca8ed62bc5"} Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.914135 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9731da0902b6256138ec7838b371ebb76e13b07484c4592648c0effea0a7330b"} Oct 03 12:51:05 crc kubenswrapper[4578]: I1003 12:51:05.915340 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"34ae8f417e92fbd24bdf81fe407c67c4069ecbf7aad2db6d0d6a49c236dacc63"} Oct 03 12:51:05 crc kubenswrapper[4578]: W1003 12:51:05.956233 4578 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:05 crc kubenswrapper[4578]: E1003 12:51:05.956338 4578 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.165:6443: connect: connection refused" logger="UnhandledError" Oct 03 12:51:06 crc kubenswrapper[4578]: W1003 12:51:06.174129 4578 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:06 crc kubenswrapper[4578]: E1003 12:51:06.174243 4578 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.165:6443: connect: connection refused" logger="UnhandledError" Oct 03 12:51:06 crc kubenswrapper[4578]: E1003 12:51:06.255727 4578 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.165:6443: connect: connection refused" interval="1.6s" Oct 03 12:51:06 crc kubenswrapper[4578]: W1003 12:51:06.418421 4578 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:06 crc kubenswrapper[4578]: E1003 12:51:06.418546 4578 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.165:6443: connect: connection refused" logger="UnhandledError" Oct 03 12:51:06 crc kubenswrapper[4578]: W1003 12:51:06.436031 4578 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:06 crc kubenswrapper[4578]: E1003 12:51:06.436084 4578 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.165:6443: connect: connection refused" logger="UnhandledError" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.498786 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.500377 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.500433 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.500453 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.500503 4578 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 12:51:06 crc kubenswrapper[4578]: E1003 12:51:06.501604 4578 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.165:6443: connect: connection refused" node="crc" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.837537 4578 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.919489 4578 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7d4204206079bd1a7dbc4c6dea30c0cbdfbea2e19d2e3cdaa3832bb3d6adacfb" exitCode=0 Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.919557 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7d4204206079bd1a7dbc4c6dea30c0cbdfbea2e19d2e3cdaa3832bb3d6adacfb"} Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.919690 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.920738 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.920770 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.920779 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.921656 4578 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="60be5924c0b93bbd78ef07fc29492fd17b85c06407edcecfb7763032326000b3" exitCode=0 Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.921720 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.921736 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"60be5924c0b93bbd78ef07fc29492fd17b85c06407edcecfb7763032326000b3"} Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.922782 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.922809 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.922821 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.925425 4578 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb" exitCode=0 Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.925553 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.925781 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb"} Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.926210 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.926235 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.926243 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.927171 4578 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d" exitCode=0 Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.927237 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d"} Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.927365 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.928153 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.928177 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.928187 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.929780 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c39656adf29002ba3e5c419f80a8cf68c8f7ddf9245ba004d0aa93930fbabece"} Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.929805 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"eb906d24f8d6787738a916592c2d2aed500fa5b1b5f7366b2f836e9a69d18e7b"} Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.957654 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.958659 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.958688 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:06 crc kubenswrapper[4578]: I1003 12:51:06.958700 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.836439 4578 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:07 crc kubenswrapper[4578]: E1003 12:51:07.861750 4578 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.165:6443: connect: connection refused" interval="3.2s" Oct 03 12:51:07 crc kubenswrapper[4578]: W1003 12:51:07.931524 4578 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:07 crc kubenswrapper[4578]: E1003 12:51:07.931616 4578 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.165:6443: connect: connection refused" logger="UnhandledError" Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.934171 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"40f9f5d7ef3a969125784cd3618c25152b78d9f8f554241f1b69d24f317ee2ef"} Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.934237 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d63272119a51846872a26ee010f0edb31fbafdb189de6425227d750de1634d76"} Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.934351 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.936051 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.936081 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.936091 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.938040 4578 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="69c39d03a71e9b2566a1b59d75b22956ddd2c505fcfe381cae908d50a0590ef8" exitCode=0 Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.938095 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"69c39d03a71e9b2566a1b59d75b22956ddd2c505fcfe381cae908d50a0590ef8"} Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.938209 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.939135 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.939154 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.939164 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.941024 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a5ede2cac008ac6d78a301b36ea4cdd25ccc11646c529d397c230d264590fcbf"} Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.941099 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.941785 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.941800 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.941808 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.944022 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b6d4f0ede4a9f9262ffe08a7abc4366faf1001b25f3a6a35dfac42c3d8362030"} Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.946181 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294"} Oct 03 12:51:07 crc kubenswrapper[4578]: I1003 12:51:07.946205 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af"} Oct 03 12:51:07 crc kubenswrapper[4578]: W1003 12:51:07.948207 4578 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:07 crc kubenswrapper[4578]: E1003 12:51:07.948275 4578 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.165:6443: connect: connection refused" logger="UnhandledError" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.101805 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.103728 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.103780 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.103794 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.103833 4578 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 12:51:08 crc kubenswrapper[4578]: E1003 12:51:08.104677 4578 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.165:6443: connect: connection refused" node="crc" Oct 03 12:51:08 crc kubenswrapper[4578]: W1003 12:51:08.128473 4578 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:08 crc kubenswrapper[4578]: E1003 12:51:08.128563 4578 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.165:6443: connect: connection refused" logger="UnhandledError" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.837261 4578 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.951810 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9d1fb8c53aa23e2c70af59f270c04d09599d79718098c17a809ace5f72797303"} Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.951899 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3ba8c08b2dceaf712d8d6b27f5c3a4b6fd3a204fc494dfb8ec33ed7ce8bc16e4"} Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.951925 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.953127 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.953156 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.953167 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.957357 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5b15fbf21ea7585fbed85430d2854584002f27faef455df679ba318ae73f56b4"} Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.957402 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf"} Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.957415 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc"} Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.957444 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.958793 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.958836 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.958858 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.963438 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.963761 4578 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="47366a9654db257f55811ac4484b3f86532df669c562b68f597b938f322d916f" exitCode=0 Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.963922 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"47366a9654db257f55811ac4484b3f86532df669c562b68f597b938f322d916f"} Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.963936 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.964038 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.964598 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.964617 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.964625 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.964749 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.964781 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.964797 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.965685 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.965705 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:08 crc kubenswrapper[4578]: I1003 12:51:08.965717 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:09 crc kubenswrapper[4578]: I1003 12:51:09.072436 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 12:51:09 crc kubenswrapper[4578]: I1003 12:51:09.097944 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 12:51:09 crc kubenswrapper[4578]: W1003 12:51:09.190939 4578 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:09 crc kubenswrapper[4578]: E1003 12:51:09.191066 4578 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.165:6443: connect: connection refused" logger="UnhandledError" Oct 03 12:51:09 crc kubenswrapper[4578]: I1003 12:51:09.837508 4578 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:09 crc kubenswrapper[4578]: I1003 12:51:09.971070 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"49927d3d11de7a2eff961d5cc4a6805a23dee422fa83f2c5ddd884ca8d2de0df"} Oct 03 12:51:09 crc kubenswrapper[4578]: I1003 12:51:09.971301 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:09 crc kubenswrapper[4578]: I1003 12:51:09.971354 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 12:51:09 crc kubenswrapper[4578]: I1003 12:51:09.971778 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:09 crc kubenswrapper[4578]: I1003 12:51:09.972789 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:09 crc kubenswrapper[4578]: I1003 12:51:09.972848 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:09 crc kubenswrapper[4578]: I1003 12:51:09.972866 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:09 crc kubenswrapper[4578]: I1003 12:51:09.974255 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:09 crc kubenswrapper[4578]: I1003 12:51:09.974319 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:09 crc kubenswrapper[4578]: I1003 12:51:09.974347 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:10 crc kubenswrapper[4578]: I1003 12:51:10.658056 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 12:51:10 crc kubenswrapper[4578]: I1003 12:51:10.658910 4578 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Oct 03 12:51:10 crc kubenswrapper[4578]: I1003 12:51:10.659085 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": dial tcp 192.168.126.11:6443: connect: connection refused" Oct 03 12:51:10 crc kubenswrapper[4578]: I1003 12:51:10.838002 4578 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:10 crc kubenswrapper[4578]: I1003 12:51:10.978424 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"780ac093760f0fec76eae4476b186e59dae84112d66c908fd1a023c235fcd8e2"} Oct 03 12:51:10 crc kubenswrapper[4578]: I1003 12:51:10.978610 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:10 crc kubenswrapper[4578]: I1003 12:51:10.978610 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:10 crc kubenswrapper[4578]: I1003 12:51:10.980374 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:10 crc kubenswrapper[4578]: I1003 12:51:10.980420 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:10 crc kubenswrapper[4578]: I1003 12:51:10.980425 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:10 crc kubenswrapper[4578]: I1003 12:51:10.980464 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:10 crc kubenswrapper[4578]: I1003 12:51:10.980486 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:10 crc kubenswrapper[4578]: I1003 12:51:10.980436 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:11 crc kubenswrapper[4578]: E1003 12:51:11.063474 4578 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.165:6443: connect: connection refused" interval="6.4s" Oct 03 12:51:11 crc kubenswrapper[4578]: I1003 12:51:11.305777 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:11 crc kubenswrapper[4578]: I1003 12:51:11.307331 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:11 crc kubenswrapper[4578]: I1003 12:51:11.307384 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:11 crc kubenswrapper[4578]: I1003 12:51:11.307399 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:11 crc kubenswrapper[4578]: I1003 12:51:11.307431 4578 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 12:51:11 crc kubenswrapper[4578]: E1003 12:51:11.308212 4578 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.165:6443: connect: connection refused" node="crc" Oct 03 12:51:11 crc kubenswrapper[4578]: I1003 12:51:11.622204 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 12:51:11 crc kubenswrapper[4578]: I1003 12:51:11.622394 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:11 crc kubenswrapper[4578]: I1003 12:51:11.623873 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:11 crc kubenswrapper[4578]: I1003 12:51:11.623937 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:11 crc kubenswrapper[4578]: I1003 12:51:11.623954 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:11 crc kubenswrapper[4578]: I1003 12:51:11.840576 4578 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:11 crc kubenswrapper[4578]: I1003 12:51:11.984690 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 03 12:51:11 crc kubenswrapper[4578]: I1003 12:51:11.987906 4578 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5b15fbf21ea7585fbed85430d2854584002f27faef455df679ba318ae73f56b4" exitCode=255 Oct 03 12:51:11 crc kubenswrapper[4578]: I1003 12:51:11.987964 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"5b15fbf21ea7585fbed85430d2854584002f27faef455df679ba318ae73f56b4"} Oct 03 12:51:11 crc kubenswrapper[4578]: I1003 12:51:11.988135 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:11 crc kubenswrapper[4578]: I1003 12:51:11.988985 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:11 crc kubenswrapper[4578]: I1003 12:51:11.989012 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:11 crc kubenswrapper[4578]: I1003 12:51:11.989022 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:11 crc kubenswrapper[4578]: I1003 12:51:11.989549 4578 scope.go:117] "RemoveContainer" containerID="5b15fbf21ea7585fbed85430d2854584002f27faef455df679ba318ae73f56b4" Oct 03 12:51:12 crc kubenswrapper[4578]: I1003 12:51:12.137885 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 12:51:12 crc kubenswrapper[4578]: I1003 12:51:12.138201 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:12 crc kubenswrapper[4578]: I1003 12:51:12.139866 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:12 crc kubenswrapper[4578]: I1003 12:51:12.139948 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:12 crc kubenswrapper[4578]: I1003 12:51:12.139964 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:12 crc kubenswrapper[4578]: W1003 12:51:12.784643 4578 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:12 crc kubenswrapper[4578]: E1003 12:51:12.784747 4578 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.165:6443: connect: connection refused" logger="UnhandledError" Oct 03 12:51:12 crc kubenswrapper[4578]: I1003 12:51:12.837442 4578 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:12 crc kubenswrapper[4578]: I1003 12:51:12.993912 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6b89889e71c0f28841554cbe4259d4c6e176a5e2cfd7704c046cabde97620444"} Oct 03 12:51:13 crc kubenswrapper[4578]: W1003 12:51:13.656698 4578 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.165:6443: connect: connection refused Oct 03 12:51:13 crc kubenswrapper[4578]: E1003 12:51:13.656828 4578 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.165:6443: connect: connection refused" logger="UnhandledError" Oct 03 12:51:14 crc kubenswrapper[4578]: I1003 12:51:14.013548 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 03 12:51:14 crc kubenswrapper[4578]: I1003 12:51:14.016557 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1"} Oct 03 12:51:14 crc kubenswrapper[4578]: I1003 12:51:14.016764 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:14 crc kubenswrapper[4578]: I1003 12:51:14.017832 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:14 crc kubenswrapper[4578]: I1003 12:51:14.017918 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:14 crc kubenswrapper[4578]: I1003 12:51:14.017975 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:14 crc kubenswrapper[4578]: I1003 12:51:14.021983 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ce12d0f8a591541a8bdf8c68c9dd153e21e01dfd6729a99e2a7227ae12a24285"} Oct 03 12:51:14 crc kubenswrapper[4578]: I1003 12:51:14.022233 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d7fed25a1a59a4c01360e4a4d5a48902dadba7455d77cbcc5905916972f8898e"} Oct 03 12:51:14 crc kubenswrapper[4578]: I1003 12:51:14.022401 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:14 crc kubenswrapper[4578]: I1003 12:51:14.023260 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:14 crc kubenswrapper[4578]: I1003 12:51:14.023377 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:14 crc kubenswrapper[4578]: I1003 12:51:14.023477 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:14 crc kubenswrapper[4578]: I1003 12:51:14.736284 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 12:51:14 crc kubenswrapper[4578]: I1003 12:51:14.736621 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:14 crc kubenswrapper[4578]: I1003 12:51:14.738153 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:14 crc kubenswrapper[4578]: I1003 12:51:14.738209 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:14 crc kubenswrapper[4578]: I1003 12:51:14.738230 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:14 crc kubenswrapper[4578]: I1003 12:51:14.738617 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 03 12:51:14 crc kubenswrapper[4578]: I1003 12:51:14.757507 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 12:51:14 crc kubenswrapper[4578]: E1003 12:51:14.999570 4578 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 03 12:51:15 crc kubenswrapper[4578]: I1003 12:51:15.026207 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:15 crc kubenswrapper[4578]: I1003 12:51:15.026251 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:15 crc kubenswrapper[4578]: I1003 12:51:15.026207 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:15 crc kubenswrapper[4578]: I1003 12:51:15.026454 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 12:51:15 crc kubenswrapper[4578]: I1003 12:51:15.027687 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:15 crc kubenswrapper[4578]: I1003 12:51:15.027757 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:15 crc kubenswrapper[4578]: I1003 12:51:15.027905 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:15 crc kubenswrapper[4578]: I1003 12:51:15.027919 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:15 crc kubenswrapper[4578]: I1003 12:51:15.027868 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:15 crc kubenswrapper[4578]: I1003 12:51:15.028011 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:15 crc kubenswrapper[4578]: I1003 12:51:15.029116 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:15 crc kubenswrapper[4578]: I1003 12:51:15.029164 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:15 crc kubenswrapper[4578]: I1003 12:51:15.029178 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:16 crc kubenswrapper[4578]: I1003 12:51:16.029260 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:16 crc kubenswrapper[4578]: I1003 12:51:16.029273 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:16 crc kubenswrapper[4578]: I1003 12:51:16.030685 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:16 crc kubenswrapper[4578]: I1003 12:51:16.030741 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:16 crc kubenswrapper[4578]: I1003 12:51:16.030757 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:16 crc kubenswrapper[4578]: I1003 12:51:16.031474 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:16 crc kubenswrapper[4578]: I1003 12:51:16.031517 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:16 crc kubenswrapper[4578]: I1003 12:51:16.031529 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:17 crc kubenswrapper[4578]: I1003 12:51:17.513119 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 12:51:17 crc kubenswrapper[4578]: I1003 12:51:17.513458 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:17 crc kubenswrapper[4578]: I1003 12:51:17.515327 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:17 crc kubenswrapper[4578]: I1003 12:51:17.515404 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:17 crc kubenswrapper[4578]: I1003 12:51:17.515425 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:17 crc kubenswrapper[4578]: I1003 12:51:17.518088 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 12:51:17 crc kubenswrapper[4578]: I1003 12:51:17.709167 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:17 crc kubenswrapper[4578]: I1003 12:51:17.711436 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:17 crc kubenswrapper[4578]: I1003 12:51:17.711490 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:17 crc kubenswrapper[4578]: I1003 12:51:17.711507 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:17 crc kubenswrapper[4578]: I1003 12:51:17.711542 4578 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 12:51:18 crc kubenswrapper[4578]: I1003 12:51:18.035288 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:18 crc kubenswrapper[4578]: I1003 12:51:18.037116 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:18 crc kubenswrapper[4578]: I1003 12:51:18.037184 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:18 crc kubenswrapper[4578]: I1003 12:51:18.037214 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:20 crc kubenswrapper[4578]: I1003 12:51:20.368189 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 03 12:51:20 crc kubenswrapper[4578]: I1003 12:51:20.368366 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:20 crc kubenswrapper[4578]: I1003 12:51:20.369712 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:20 crc kubenswrapper[4578]: I1003 12:51:20.369746 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:20 crc kubenswrapper[4578]: I1003 12:51:20.369755 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:20 crc kubenswrapper[4578]: I1003 12:51:20.513953 4578 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 03 12:51:20 crc kubenswrapper[4578]: I1003 12:51:20.514056 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 03 12:51:20 crc kubenswrapper[4578]: I1003 12:51:20.871981 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 03 12:51:21 crc kubenswrapper[4578]: I1003 12:51:21.042583 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:21 crc kubenswrapper[4578]: I1003 12:51:21.043705 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:21 crc kubenswrapper[4578]: I1003 12:51:21.043776 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:21 crc kubenswrapper[4578]: I1003 12:51:21.043788 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:21 crc kubenswrapper[4578]: I1003 12:51:21.063131 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 03 12:51:22 crc kubenswrapper[4578]: I1003 12:51:22.043963 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:22 crc kubenswrapper[4578]: I1003 12:51:22.048538 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:22 crc kubenswrapper[4578]: I1003 12:51:22.048915 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:22 crc kubenswrapper[4578]: I1003 12:51:22.048949 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:23 crc kubenswrapper[4578]: I1003 12:51:23.836873 4578 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 03 12:51:24 crc kubenswrapper[4578]: W1003 12:51:24.232561 4578 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 03 12:51:24 crc kubenswrapper[4578]: I1003 12:51:24.232681 4578 trace.go:236] Trace[1182498211]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Oct-2025 12:51:14.229) (total time: 10003ms): Oct 03 12:51:24 crc kubenswrapper[4578]: Trace[1182498211]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10003ms (12:51:24.232) Oct 03 12:51:24 crc kubenswrapper[4578]: Trace[1182498211]: [10.003157908s] [10.003157908s] END Oct 03 12:51:24 crc kubenswrapper[4578]: E1003 12:51:24.232703 4578 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 03 12:51:24 crc kubenswrapper[4578]: E1003 12:51:24.285656 4578 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.186afc2ab7298973 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-03 12:51:04.833378675 +0000 UTC m=+0.631850899,LastTimestamp:2025-10-03 12:51:04.833378675 +0000 UTC m=+0.631850899,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 03 12:51:24 crc kubenswrapper[4578]: I1003 12:51:24.559756 4578 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 03 12:51:24 crc kubenswrapper[4578]: I1003 12:51:24.559834 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 03 12:51:24 crc kubenswrapper[4578]: I1003 12:51:24.569153 4578 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 03 12:51:24 crc kubenswrapper[4578]: I1003 12:51:24.569216 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 03 12:51:25 crc kubenswrapper[4578]: E1003 12:51:25.000283 4578 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 03 12:51:25 crc kubenswrapper[4578]: I1003 12:51:25.661944 4578 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]log ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]etcd ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/openshift.io-api-request-count-filter ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/openshift.io-startkubeinformers ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/generic-apiserver-start-informers ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/priority-and-fairness-config-consumer ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/priority-and-fairness-filter ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/start-apiextensions-informers ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/start-apiextensions-controllers ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/crd-informer-synced ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/start-system-namespaces-controller ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/start-cluster-authentication-info-controller ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/start-legacy-token-tracking-controller ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/start-service-ip-repair-controllers ok Oct 03 12:51:25 crc kubenswrapper[4578]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/priority-and-fairness-config-producer ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/bootstrap-controller ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/start-kube-aggregator-informers ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/apiservice-status-local-available-controller ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/apiservice-status-remote-available-controller ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/apiservice-registration-controller ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/apiservice-wait-for-first-sync ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/apiservice-discovery-controller ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/kube-apiserver-autoregistration ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]autoregister-completion ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/apiservice-openapi-controller ok Oct 03 12:51:25 crc kubenswrapper[4578]: [+]poststarthook/apiservice-openapiv3-controller ok Oct 03 12:51:25 crc kubenswrapper[4578]: livez check failed Oct 03 12:51:25 crc kubenswrapper[4578]: I1003 12:51:25.662003 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 12:51:29 crc kubenswrapper[4578]: E1003 12:51:29.540546 4578 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="7s" Oct 03 12:51:29 crc kubenswrapper[4578]: I1003 12:51:29.542542 4578 trace.go:236] Trace[829150838]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Oct-2025 12:51:14.948) (total time: 14594ms): Oct 03 12:51:29 crc kubenswrapper[4578]: Trace[829150838]: ---"Objects listed" error: 14594ms (12:51:29.542) Oct 03 12:51:29 crc kubenswrapper[4578]: Trace[829150838]: [14.594199463s] [14.594199463s] END Oct 03 12:51:29 crc kubenswrapper[4578]: I1003 12:51:29.542601 4578 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 03 12:51:29 crc kubenswrapper[4578]: I1003 12:51:29.544292 4578 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 03 12:51:29 crc kubenswrapper[4578]: E1003 12:51:29.544715 4578 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 03 12:51:29 crc kubenswrapper[4578]: I1003 12:51:29.546813 4578 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 03 12:51:29 crc kubenswrapper[4578]: I1003 12:51:29.549482 4578 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 03 12:51:29 crc kubenswrapper[4578]: I1003 12:51:29.575240 4578 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48778->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 03 12:51:29 crc kubenswrapper[4578]: I1003 12:51:29.575306 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48778->192.168.126.11:17697: read: connection reset by peer" Oct 03 12:51:29 crc kubenswrapper[4578]: I1003 12:51:29.575345 4578 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48790->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 03 12:51:29 crc kubenswrapper[4578]: I1003 12:51:29.575504 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48790->192.168.126.11:17697: read: connection reset by peer" Oct 03 12:51:30 crc kubenswrapper[4578]: I1003 12:51:30.063260 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 03 12:51:30 crc kubenswrapper[4578]: I1003 12:51:30.064338 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 03 12:51:30 crc kubenswrapper[4578]: I1003 12:51:30.066094 4578 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1" exitCode=255 Oct 03 12:51:30 crc kubenswrapper[4578]: I1003 12:51:30.066144 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1"} Oct 03 12:51:30 crc kubenswrapper[4578]: I1003 12:51:30.066192 4578 scope.go:117] "RemoveContainer" containerID="5b15fbf21ea7585fbed85430d2854584002f27faef455df679ba318ae73f56b4" Oct 03 12:51:30 crc kubenswrapper[4578]: I1003 12:51:30.066326 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:30 crc kubenswrapper[4578]: I1003 12:51:30.067926 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:30 crc kubenswrapper[4578]: I1003 12:51:30.068023 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:30 crc kubenswrapper[4578]: I1003 12:51:30.068103 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:30 crc kubenswrapper[4578]: I1003 12:51:30.068936 4578 scope.go:117] "RemoveContainer" containerID="9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1" Oct 03 12:51:30 crc kubenswrapper[4578]: E1003 12:51:30.069258 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 03 12:51:30 crc kubenswrapper[4578]: I1003 12:51:30.341976 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 12:51:30 crc kubenswrapper[4578]: I1003 12:51:30.342174 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:30 crc kubenswrapper[4578]: I1003 12:51:30.343359 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:30 crc kubenswrapper[4578]: I1003 12:51:30.343508 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:30 crc kubenswrapper[4578]: I1003 12:51:30.343597 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:30 crc kubenswrapper[4578]: I1003 12:51:30.346078 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 03 12:51:30 crc kubenswrapper[4578]: I1003 12:51:30.662710 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.069201 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.071206 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.071273 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.072324 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.072331 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.072355 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.072371 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.072380 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.072381 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.072919 4578 scope.go:117] "RemoveContainer" containerID="9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1" Oct 03 12:51:31 crc kubenswrapper[4578]: E1003 12:51:31.073075 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.076349 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.134482 4578 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.840050 4578 apiserver.go:52] "Watching apiserver" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.848347 4578 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.848684 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-p8d2v","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-multus/multus-additional-cni-plugins-jjvmh","openshift-multus/multus-c4jgn","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-operator/iptables-alerter-4ln5h"] Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.849003 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.849089 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.849112 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.849127 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.849463 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 12:51:31 crc kubenswrapper[4578]: E1003 12:51:31.849549 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.849809 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:31 crc kubenswrapper[4578]: E1003 12:51:31.849815 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:51:31 crc kubenswrapper[4578]: E1003 12:51:31.849872 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.850152 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-p8d2v" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.850495 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-c4jgn" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.850616 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.854997 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.855252 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.855402 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.855534 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.855708 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.855836 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.855953 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.856106 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.856258 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.856751 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.856828 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.862466 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.862719 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.863102 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.865481 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.865564 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.865953 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.866101 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.867386 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.872884 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.886189 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.899355 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.913439 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.923752 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.933642 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.944871 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.948836 4578 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.955092 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.962056 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.962095 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.962141 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.962158 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.962373 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.962436 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.962435 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.962179 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963230 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963249 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963265 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963270 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963280 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963344 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963366 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963383 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963399 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963415 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963429 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963444 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963456 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963485 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963508 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963501 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963530 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963548 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963567 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963597 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963612 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963649 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963665 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963707 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963723 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963736 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963752 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963766 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963780 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963810 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963825 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963856 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963873 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963889 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963904 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963919 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963933 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963947 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963963 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963978 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963995 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964012 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964026 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964040 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964074 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964088 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964104 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964117 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964142 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964156 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964171 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964185 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964202 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964282 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.963666 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964034 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964429 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964581 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: E1003 12:51:31.964592 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:51:32.464571628 +0000 UTC m=+28.263043812 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965286 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965316 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965334 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965354 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965373 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965389 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965409 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965424 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965439 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965454 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965471 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965514 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965529 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965546 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965562 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965578 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965593 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965612 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965652 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965670 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965693 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965755 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965776 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965793 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965814 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965855 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965873 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965892 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965909 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965925 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965943 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965960 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965980 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965996 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966013 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966030 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966046 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966064 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966082 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966100 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965330 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965342 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964765 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964804 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964905 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964917 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965007 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965061 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965142 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965187 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965201 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966211 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965216 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965422 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965554 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965601 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965611 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965762 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965852 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965906 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.965995 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966068 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966110 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966192 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966280 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966361 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966397 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966428 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966469 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966496 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966565 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966605 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.964751 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966673 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966756 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966774 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966118 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966812 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966830 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966839 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966847 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966863 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966879 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966895 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966911 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966927 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966942 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966958 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966978 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967011 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967018 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967040 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967065 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967090 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967107 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967123 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967139 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967154 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967169 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967186 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967201 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967201 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967200 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967217 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967235 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967257 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967273 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967288 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967303 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967318 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967332 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967341 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967346 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967386 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967390 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967413 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967436 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967460 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967464 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967484 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967506 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967514 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967528 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967552 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967573 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967597 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967621 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967661 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967670 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967683 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967705 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967728 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967755 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967780 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967804 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967826 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967848 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967870 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967890 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967912 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967944 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967964 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.967985 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968011 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968031 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968049 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968143 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968166 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968189 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968211 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968235 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968258 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968280 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968299 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968319 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968356 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968380 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968403 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968424 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968446 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968468 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968489 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968511 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968531 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968554 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968574 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968595 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968617 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968658 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968683 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968706 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968727 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968748 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968770 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968794 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968816 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968837 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968861 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968891 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968917 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968943 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968974 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.968973 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969001 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969027 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969088 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969115 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969139 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969165 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969196 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5403769c-f73e-4921-a402-3d650774c2e7-cnibin\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969220 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5403769c-f73e-4921-a402-3d650774c2e7-cni-binary-copy\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969244 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969265 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-cnibin\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969285 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fed3968f-4bb0-4209-a377-16c26f9ea070-cni-binary-copy\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969309 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-etc-kubernetes\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969331 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5403769c-f73e-4921-a402-3d650774c2e7-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969353 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-hostroot\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969374 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-multus-conf-dir\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969395 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba-hosts-file\") pod \"node-resolver-p8d2v\" (UID: \"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\") " pod="openshift-dns/node-resolver-p8d2v" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969417 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5403769c-f73e-4921-a402-3d650774c2e7-os-release\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969442 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969464 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-host-var-lib-kubelet\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969509 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-host-run-netns\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969533 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxn6b\" (UniqueName: \"kubernetes.io/projected/fed3968f-4bb0-4209-a377-16c26f9ea070-kube-api-access-qxn6b\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969556 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5403769c-f73e-4921-a402-3d650774c2e7-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969582 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969608 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969650 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5403769c-f73e-4921-a402-3d650774c2e7-system-cni-dir\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969673 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969695 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-host-run-k8s-cni-cncf-io\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969749 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969753 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969774 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-system-cni-dir\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969834 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969859 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969874 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969886 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969916 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-multus-socket-dir-parent\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969938 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrtvm\" (UniqueName: \"kubernetes.io/projected/5403769c-f73e-4921-a402-3d650774c2e7-kube-api-access-rrtvm\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969959 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969981 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-host-var-lib-cni-multus\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969999 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-host-run-multus-certs\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.969998 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970017 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-os-release\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970035 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-host-var-lib-cni-bin\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970053 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvxhl\" (UniqueName: \"kubernetes.io/projected/fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba-kube-api-access-qvxhl\") pod \"node-resolver-p8d2v\" (UID: \"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\") " pod="openshift-dns/node-resolver-p8d2v" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970070 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-multus-cni-dir\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970080 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970048 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970110 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970361 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970366 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970396 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970445 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970086 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/fed3968f-4bb0-4209-a377-16c26f9ea070-multus-daemon-config\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970475 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970488 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.966270 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8qvh9"] Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970601 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970613 4578 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970648 4578 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970648 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970669 4578 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970685 4578 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970703 4578 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970717 4578 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970730 4578 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970746 4578 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970775 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970710 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970790 4578 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970819 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970831 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970843 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970854 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970865 4578 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970881 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970893 4578 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970903 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970904 4578 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970949 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970960 4578 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970971 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970986 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.970997 4578 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971007 4578 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971009 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971018 4578 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971041 4578 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971061 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971071 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971082 4578 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971093 4578 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971104 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971114 4578 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971124 4578 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971134 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971265 4578 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971262 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971273 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971278 4578 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971304 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971352 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971362 4578 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971354 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971379 4578 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971470 4578 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971488 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971505 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971521 4578 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971535 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971548 4578 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971562 4578 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971575 4578 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971589 4578 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971602 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971616 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971646 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971662 4578 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971679 4578 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971693 4578 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971706 4578 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971721 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971735 4578 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971751 4578 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971769 4578 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971788 4578 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971487 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971712 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.971729 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.972020 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.972072 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.972112 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.972197 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.972302 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.972446 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.972789 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.978347 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.980190 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.980703 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.981312 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.981441 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.981298 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.981680 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.981866 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.982136 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.982147 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.982331 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.982502 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.982708 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.982841 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.982921 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.983110 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.989295 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.989918 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.990548 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.990749 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.990795 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.991066 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.991238 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: E1003 12:51:31.991255 4578 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 12:51:31 crc kubenswrapper[4578]: E1003 12:51:31.991357 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:32.491315517 +0000 UTC m=+28.289787771 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.991489 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.991827 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.991839 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.992488 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.994664 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: E1003 12:51:31.996299 4578 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.996500 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: E1003 12:51:31.996602 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:32.496506409 +0000 UTC m=+28.294978593 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.997515 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.998443 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.998682 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.998813 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.998975 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 12:51:31 crc kubenswrapper[4578]: I1003 12:51:31.999170 4578 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.003086 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.004141 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.004726 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.005039 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.005278 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.005670 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.005978 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.006146 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.006329 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.006554 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.006818 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.006895 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.006915 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.007186 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.007396 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.007603 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.007649 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.007835 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.007921 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.008009 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.008020 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.008208 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.008273 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.008464 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.008664 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.008862 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.008881 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.009305 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.009481 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.011122 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.011259 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.011436 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.011703 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.012097 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.012287 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.012375 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.012817 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.014512 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.014951 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.015220 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.015112 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.015138 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.015280 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.012822 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.012998 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.015348 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.015423 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.015860 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.015923 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.013116 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.016398 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.016503 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.016523 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.016518 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.016536 4578 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.016619 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:32.51659296 +0000 UTC m=+28.315065144 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.016977 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.017519 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.017605 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.017861 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:31.971394 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-qq9jm"] Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.018037 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.018366 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.018685 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.018941 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.019534 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.019548 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.019933 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.024824 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:31.971471 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.025316 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.025469 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.027392 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.027545 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.028909 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.029063 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.029327 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.029475 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.029646 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.031747 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.031975 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.034426 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.034548 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.034654 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.034776 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.035433 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.037759 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.037911 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.038337 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.038548 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.038612 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.040108 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.040348 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.040374 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.040388 4578 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.040462 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:32.540444419 +0000 UTC m=+28.338916603 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.040694 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.041378 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.041702 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.041924 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.043787 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.045793 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.045908 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.052430 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.060197 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.061018 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.065696 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.071958 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.072833 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b7d7e663-26ba-48e2-849e-f7c9554fd7a7-rootfs\") pod \"machine-config-daemon-qq9jm\" (UID: \"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\") " pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.072861 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b7d7e663-26ba-48e2-849e-f7c9554fd7a7-mcd-auth-proxy-config\") pod \"machine-config-daemon-qq9jm\" (UID: \"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\") " pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.072889 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-run-openvswitch\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.072904 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.072923 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-multus-socket-dir-parent\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.072937 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-etc-openvswitch\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.072952 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-cni-netd\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.072968 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrtvm\" (UniqueName: \"kubernetes.io/projected/5403769c-f73e-4921-a402-3d650774c2e7-kube-api-access-rrtvm\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.072984 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-kubelet\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.072998 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-run-systemd\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073029 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-node-log\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073044 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-ovnkube-script-lib\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073063 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-host-var-lib-cni-multus\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073080 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-host-run-multus-certs\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073102 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-os-release\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073118 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-host-var-lib-cni-bin\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073135 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvxhl\" (UniqueName: \"kubernetes.io/projected/fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba-kube-api-access-qvxhl\") pod \"node-resolver-p8d2v\" (UID: \"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\") " pod="openshift-dns/node-resolver-p8d2v" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073151 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-multus-cni-dir\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073165 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/fed3968f-4bb0-4209-a377-16c26f9ea070-multus-daemon-config\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073192 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-var-lib-openvswitch\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073223 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-ovn-node-metrics-cert\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073246 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-cnibin\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073262 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fed3968f-4bb0-4209-a377-16c26f9ea070-cni-binary-copy\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073277 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5403769c-f73e-4921-a402-3d650774c2e7-cnibin\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073294 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5403769c-f73e-4921-a402-3d650774c2e7-cni-binary-copy\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073313 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073375 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-etc-kubernetes\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073400 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-cni-bin\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073418 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-log-socket\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073447 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5403769c-f73e-4921-a402-3d650774c2e7-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073473 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-hostroot\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073492 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-multus-conf-dir\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073517 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-run-ovn\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073536 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-ovnkube-config\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073556 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gt6c5\" (UniqueName: \"kubernetes.io/projected/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-kube-api-access-gt6c5\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073573 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5403769c-f73e-4921-a402-3d650774c2e7-os-release\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073589 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-host-var-lib-kubelet\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073604 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba-hosts-file\") pod \"node-resolver-p8d2v\" (UID: \"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\") " pod="openshift-dns/node-resolver-p8d2v" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073666 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-host-run-netns\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073770 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkhmc\" (UniqueName: \"kubernetes.io/projected/b7d7e663-26ba-48e2-849e-f7c9554fd7a7-kube-api-access-nkhmc\") pod \"machine-config-daemon-qq9jm\" (UID: \"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\") " pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073789 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-env-overrides\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073804 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b7d7e663-26ba-48e2-849e-f7c9554fd7a7-proxy-tls\") pod \"machine-config-daemon-qq9jm\" (UID: \"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\") " pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073818 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-systemd-units\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073836 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5403769c-f73e-4921-a402-3d650774c2e7-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073867 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxn6b\" (UniqueName: \"kubernetes.io/projected/fed3968f-4bb0-4209-a377-16c26f9ea070-kube-api-access-qxn6b\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073883 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5403769c-f73e-4921-a402-3d650774c2e7-system-cni-dir\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073900 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073915 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-run-netns\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073934 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-host-run-k8s-cni-cncf-io\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073949 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-slash\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073965 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-system-cni-dir\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.073982 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-run-ovn-kubernetes\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074023 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074034 4578 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074044 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074054 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074063 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074074 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074082 4578 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074091 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074100 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074108 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074117 4578 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074126 4578 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074137 4578 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074155 4578 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074170 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074182 4578 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074195 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074209 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074238 4578 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074259 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074283 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074298 4578 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074311 4578 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074322 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074334 4578 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074346 4578 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074358 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074370 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074392 4578 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074406 4578 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074418 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074432 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074444 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074455 4578 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074467 4578 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074487 4578 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074501 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074515 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074526 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074538 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074548 4578 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074558 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074570 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074584 4578 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074617 4578 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074646 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074657 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074669 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074681 4578 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074693 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074706 4578 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074718 4578 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074731 4578 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074742 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074755 4578 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074766 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074777 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074789 4578 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074802 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074815 4578 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074846 4578 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074858 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074871 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074883 4578 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074896 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074908 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074920 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074931 4578 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074944 4578 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.074956 4578 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.075023 4578 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.075066 4578 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.075088 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.075087 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5403769c-f73e-4921-a402-3d650774c2e7-cni-binary-copy\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.075103 4578 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.075154 4578 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.075165 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.075173 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.075174 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-host-run-multus-certs\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.075203 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.075214 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-multus-socket-dir-parent\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.075223 4578 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.075259 4578 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.075260 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-host-var-lib-cni-bin\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.075488 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-system-cni-dir\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.075519 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5403769c-f73e-4921-a402-3d650774c2e7-system-cni-dir\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.075549 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-cnibin\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.075571 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-host-var-lib-kubelet\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.075725 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5403769c-f73e-4921-a402-3d650774c2e7-os-release\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.076332 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fed3968f-4bb0-4209-a377-16c26f9ea070-cni-binary-copy\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.076384 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba-hosts-file\") pod \"node-resolver-p8d2v\" (UID: \"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\") " pod="openshift-dns/node-resolver-p8d2v" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.076402 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5403769c-f73e-4921-a402-3d650774c2e7-cnibin\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.076420 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-host-run-netns\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.076559 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5403769c-f73e-4921-a402-3d650774c2e7-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.076614 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-host-run-k8s-cni-cncf-io\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.076718 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.076768 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-multus-conf-dir\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.076814 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-os-release\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.076843 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-host-var-lib-cni-multus\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.076898 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.076855 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-hostroot\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.076916 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-multus-cni-dir\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.076924 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fed3968f-4bb0-4209-a377-16c26f9ea070-etc-kubernetes\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077140 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077179 4578 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077197 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077209 4578 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077223 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077234 4578 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077248 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077406 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077424 4578 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077435 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077446 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077462 4578 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077473 4578 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077484 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077495 4578 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077506 4578 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077517 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077527 4578 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077438 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077539 4578 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077550 4578 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077561 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077571 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077582 4578 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077593 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077603 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077613 4578 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077523 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5403769c-f73e-4921-a402-3d650774c2e7-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077623 4578 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077708 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077742 4578 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077768 4578 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077784 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077803 4578 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077818 4578 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077835 4578 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077848 4578 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077860 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077873 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077884 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077898 4578 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077908 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077917 4578 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077927 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077938 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.078001 4578 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.078017 4578 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.078029 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.078044 4578 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.078056 4578 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.078065 4578 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.078075 4578 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.078086 4578 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.078098 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.077652 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/fed3968f-4bb0-4209-a377-16c26f9ea070-multus-daemon-config\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.099807 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.100377 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxn6b\" (UniqueName: \"kubernetes.io/projected/fed3968f-4bb0-4209-a377-16c26f9ea070-kube-api-access-qxn6b\") pod \"multus-c4jgn\" (UID: \"fed3968f-4bb0-4209-a377-16c26f9ea070\") " pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.101094 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvxhl\" (UniqueName: \"kubernetes.io/projected/fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba-kube-api-access-qvxhl\") pod \"node-resolver-p8d2v\" (UID: \"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\") " pod="openshift-dns/node-resolver-p8d2v" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.101294 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrtvm\" (UniqueName: \"kubernetes.io/projected/5403769c-f73e-4921-a402-3d650774c2e7-kube-api-access-rrtvm\") pod \"multus-additional-cni-plugins-jjvmh\" (UID: \"5403769c-f73e-4921-a402-3d650774c2e7\") " pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.107255 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.107536 4578 scope.go:117] "RemoveContainer" containerID="9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1" Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.107697 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.107826 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.115350 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.127574 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.139495 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.156404 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.168303 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.169255 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.175450 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.179750 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-cni-bin\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.179816 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-log-socket\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.179863 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-run-ovn\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.179887 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-ovnkube-config\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.179908 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gt6c5\" (UniqueName: \"kubernetes.io/projected/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-kube-api-access-gt6c5\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.179953 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkhmc\" (UniqueName: \"kubernetes.io/projected/b7d7e663-26ba-48e2-849e-f7c9554fd7a7-kube-api-access-nkhmc\") pod \"machine-config-daemon-qq9jm\" (UID: \"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\") " pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.179976 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-env-overrides\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.179998 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b7d7e663-26ba-48e2-849e-f7c9554fd7a7-proxy-tls\") pod \"machine-config-daemon-qq9jm\" (UID: \"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\") " pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.180042 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-systemd-units\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.180079 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-run-netns\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.180127 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-slash\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.180147 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-run-ovn-kubernetes\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.180187 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b7d7e663-26ba-48e2-849e-f7c9554fd7a7-rootfs\") pod \"machine-config-daemon-qq9jm\" (UID: \"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\") " pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.180210 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b7d7e663-26ba-48e2-849e-f7c9554fd7a7-mcd-auth-proxy-config\") pod \"machine-config-daemon-qq9jm\" (UID: \"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\") " pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.180239 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-run-openvswitch\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.180291 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.180313 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-etc-openvswitch\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.180354 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-cni-netd\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.180376 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-kubelet\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.180424 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-run-systemd\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.180449 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-node-log\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.180469 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-ovnkube-script-lib\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.180514 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-ovn-node-metrics-cert\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.180538 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-var-lib-openvswitch\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.180590 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.180672 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-var-lib-openvswitch\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.180740 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-cni-bin\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.180778 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-log-socket\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.181003 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-run-ovn\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.181754 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-systemd-units\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.182123 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-kubelet\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.182177 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-run-openvswitch\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.182213 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.182243 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-etc-openvswitch\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.182274 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-cni-netd\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.182307 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-node-log\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.182338 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-run-systemd\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.182353 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b7d7e663-26ba-48e2-849e-f7c9554fd7a7-mcd-auth-proxy-config\") pod \"machine-config-daemon-qq9jm\" (UID: \"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\") " pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.182373 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-slash\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.182408 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-run-netns\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.182418 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-run-ovn-kubernetes\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.182422 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-ovnkube-config\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.182441 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b7d7e663-26ba-48e2-849e-f7c9554fd7a7-rootfs\") pod \"machine-config-daemon-qq9jm\" (UID: \"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\") " pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.182720 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-env-overrides\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.184832 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.184857 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.185083 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b7d7e663-26ba-48e2-849e-f7c9554fd7a7-proxy-tls\") pod \"machine-config-daemon-qq9jm\" (UID: \"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\") " pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.185370 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-ovnkube-script-lib\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.186597 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-ovn-node-metrics-cert\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.191990 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-p8d2v" Oct 03 12:51:32 crc kubenswrapper[4578]: W1003 12:51:32.192909 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-8b203e9e060c7cdf501f3e7f6741565007472be37504ae7cf0827caa93d480dd WatchSource:0}: Error finding container 8b203e9e060c7cdf501f3e7f6741565007472be37504ae7cf0827caa93d480dd: Status 404 returned error can't find the container with id 8b203e9e060c7cdf501f3e7f6741565007472be37504ae7cf0827caa93d480dd Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.199659 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" Oct 03 12:51:32 crc kubenswrapper[4578]: W1003 12:51:32.204205 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-03eeb26e305a68358492053fdc323b592d54e051dcd93f8bb6219089f53c3ce2 WatchSource:0}: Error finding container 03eeb26e305a68358492053fdc323b592d54e051dcd93f8bb6219089f53c3ce2: Status 404 returned error can't find the container with id 03eeb26e305a68358492053fdc323b592d54e051dcd93f8bb6219089f53c3ce2 Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.205466 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-c4jgn" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.206759 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gt6c5\" (UniqueName: \"kubernetes.io/projected/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-kube-api-access-gt6c5\") pod \"ovnkube-node-8qvh9\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.206828 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkhmc\" (UniqueName: \"kubernetes.io/projected/b7d7e663-26ba-48e2-849e-f7c9554fd7a7-kube-api-access-nkhmc\") pod \"machine-config-daemon-qq9jm\" (UID: \"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\") " pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.208434 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.220363 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.246849 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: W1003 12:51:32.252797 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5403769c_f73e_4921_a402_3d650774c2e7.slice/crio-7754315f1abbb06716729bacdda8343eb36292155808841e7bc17dda69bc53b3 WatchSource:0}: Error finding container 7754315f1abbb06716729bacdda8343eb36292155808841e7bc17dda69bc53b3: Status 404 returned error can't find the container with id 7754315f1abbb06716729bacdda8343eb36292155808841e7bc17dda69bc53b3 Oct 03 12:51:32 crc kubenswrapper[4578]: W1003 12:51:32.254314 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfbf63262_15f9_4fe4_b6f7_a3fc69bc07ba.slice/crio-9237c32ed2e807bb340de87bdb5cf3f76b7ae775810401cd86b161efcc1cbc77 WatchSource:0}: Error finding container 9237c32ed2e807bb340de87bdb5cf3f76b7ae775810401cd86b161efcc1cbc77: Status 404 returned error can't find the container with id 9237c32ed2e807bb340de87bdb5cf3f76b7ae775810401cd86b161efcc1cbc77 Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.264961 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: W1003 12:51:32.276904 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfed3968f_4bb0_4209_a377_16c26f9ea070.slice/crio-a8064835f0d13a35eaef6cebe20ef8f5cc4284e79bbbe02ba81e51671fca1b33 WatchSource:0}: Error finding container a8064835f0d13a35eaef6cebe20ef8f5cc4284e79bbbe02ba81e51671fca1b33: Status 404 returned error can't find the container with id a8064835f0d13a35eaef6cebe20ef8f5cc4284e79bbbe02ba81e51671fca1b33 Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.282038 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.297005 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.313412 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.335253 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.348310 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.356868 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.364153 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.372774 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.392529 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: W1003 12:51:32.409759 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7d7e663_26ba_48e2_849e_f7c9554fd7a7.slice/crio-8ae1feeb5dddff6b61d02813cb70e6100f5822a6d9f35a06848608b86a8f1695 WatchSource:0}: Error finding container 8ae1feeb5dddff6b61d02813cb70e6100f5822a6d9f35a06848608b86a8f1695: Status 404 returned error can't find the container with id 8ae1feeb5dddff6b61d02813cb70e6100f5822a6d9f35a06848608b86a8f1695 Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.415670 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.456550 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.483908 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.484219 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:51:33.484193572 +0000 UTC m=+29.282665756 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.585555 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.585614 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.585662 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.585683 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.585796 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.585810 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.585820 4578 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.585860 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:33.585846822 +0000 UTC m=+29.384319006 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.585905 4578 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.585952 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:33.585920314 +0000 UTC m=+29.384392498 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.585978 4578 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.585995 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:33.585990306 +0000 UTC m=+29.384462490 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.586083 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.586093 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.586100 4578 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.586122 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:33.58611614 +0000 UTC m=+29.384588324 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.909003 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:32 crc kubenswrapper[4578]: E1003 12:51:32.909169 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.914566 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.915443 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.916191 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.916895 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.917462 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.917941 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.918544 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.919147 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.919798 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.920379 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.920969 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.921699 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.922263 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.923675 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.924929 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.925882 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.926938 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.927511 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.929475 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.930378 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.931765 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.932688 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.933482 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.935330 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.936208 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.938160 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.939269 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.940847 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.941567 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.942603 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.943243 4578 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.943349 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.946009 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.946532 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.946987 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.948994 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.950285 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.950887 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.951927 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.952768 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.953621 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.954270 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.955282 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.955918 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.956864 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.957477 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.958668 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.959449 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.960367 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.960875 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.961903 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.962557 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.963192 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 03 12:51:32 crc kubenswrapper[4578]: I1003 12:51:32.964091 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.087011 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"03eeb26e305a68358492053fdc323b592d54e051dcd93f8bb6219089f53c3ce2"} Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.088911 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff"} Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.088936 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12"} Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.088948 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"44eddc505beebd9c0b407404e2fc633918492931bd68911dd9e912d78c35abad"} Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.091588 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734"} Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.091626 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"8b203e9e060c7cdf501f3e7f6741565007472be37504ae7cf0827caa93d480dd"} Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.093569 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-c4jgn" event={"ID":"fed3968f-4bb0-4209-a377-16c26f9ea070","Type":"ContainerStarted","Data":"7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3"} Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.093649 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-c4jgn" event={"ID":"fed3968f-4bb0-4209-a377-16c26f9ea070","Type":"ContainerStarted","Data":"a8064835f0d13a35eaef6cebe20ef8f5cc4284e79bbbe02ba81e51671fca1b33"} Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.096513 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716"} Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.096545 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"8ae1feeb5dddff6b61d02813cb70e6100f5822a6d9f35a06848608b86a8f1695"} Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.099383 4578 generic.go:334] "Generic (PLEG): container finished" podID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerID="0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49" exitCode=0 Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.099464 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerDied","Data":"0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49"} Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.099493 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerStarted","Data":"cac09449164b0eee0488dbd39377414aa0d07f6a27a355af9d775b75b6e18563"} Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.101755 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" event={"ID":"5403769c-f73e-4921-a402-3d650774c2e7","Type":"ContainerStarted","Data":"838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29"} Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.101795 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" event={"ID":"5403769c-f73e-4921-a402-3d650774c2e7","Type":"ContainerStarted","Data":"7754315f1abbb06716729bacdda8343eb36292155808841e7bc17dda69bc53b3"} Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.103854 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-p8d2v" event={"ID":"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba","Type":"ContainerStarted","Data":"25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50"} Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.103904 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-p8d2v" event={"ID":"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba","Type":"ContainerStarted","Data":"9237c32ed2e807bb340de87bdb5cf3f76b7ae775810401cd86b161efcc1cbc77"} Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.104315 4578 scope.go:117] "RemoveContainer" containerID="9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1" Oct 03 12:51:33 crc kubenswrapper[4578]: E1003 12:51:33.104471 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.105486 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.116385 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.125265 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.136555 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.147835 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.158479 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.171445 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.187238 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.205346 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.218599 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.229183 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.239282 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.248600 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.258662 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.268567 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.279069 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.292973 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.307753 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.321111 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.332485 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.345704 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.367211 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.380275 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.393558 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.494853 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:51:33 crc kubenswrapper[4578]: E1003 12:51:33.495074 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:51:35.495060241 +0000 UTC m=+31.293532425 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.596132 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.596182 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.596203 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.596229 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:33 crc kubenswrapper[4578]: E1003 12:51:33.596333 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 12:51:33 crc kubenswrapper[4578]: E1003 12:51:33.596347 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 12:51:33 crc kubenswrapper[4578]: E1003 12:51:33.596357 4578 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:33 crc kubenswrapper[4578]: E1003 12:51:33.596354 4578 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 12:51:33 crc kubenswrapper[4578]: E1003 12:51:33.596334 4578 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 12:51:33 crc kubenswrapper[4578]: E1003 12:51:33.596366 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 12:51:33 crc kubenswrapper[4578]: E1003 12:51:33.596491 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 12:51:33 crc kubenswrapper[4578]: E1003 12:51:33.596505 4578 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:33 crc kubenswrapper[4578]: E1003 12:51:33.596406 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:35.59638741 +0000 UTC m=+31.394859604 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:33 crc kubenswrapper[4578]: E1003 12:51:33.596575 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:35.596557246 +0000 UTC m=+31.395029430 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 12:51:33 crc kubenswrapper[4578]: E1003 12:51:33.596588 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:35.596581426 +0000 UTC m=+31.395053610 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 12:51:33 crc kubenswrapper[4578]: E1003 12:51:33.596606 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:35.596601687 +0000 UTC m=+31.395073871 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.908862 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:33 crc kubenswrapper[4578]: E1003 12:51:33.909004 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:51:33 crc kubenswrapper[4578]: I1003 12:51:33.908852 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:33 crc kubenswrapper[4578]: E1003 12:51:33.909091 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.108831 4578 generic.go:334] "Generic (PLEG): container finished" podID="5403769c-f73e-4921-a402-3d650774c2e7" containerID="838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29" exitCode=0 Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.108951 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" event={"ID":"5403769c-f73e-4921-a402-3d650774c2e7","Type":"ContainerDied","Data":"838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29"} Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.112346 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053"} Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.116214 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerStarted","Data":"c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982"} Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.116276 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerStarted","Data":"32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236"} Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.116290 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerStarted","Data":"31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d"} Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.116299 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerStarted","Data":"39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9"} Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.116308 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerStarted","Data":"73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030"} Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.125885 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.140473 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.156235 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.170086 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.189678 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.258376 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.277165 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.291064 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.308138 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.322262 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.344436 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.357094 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.372806 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.389705 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.407653 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.430251 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.447612 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.521598 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.537413 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.551046 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.563090 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.575746 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.588333 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.601570 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.898723 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-rh9f2"] Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.902819 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-rh9f2" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.904833 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.905290 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.905562 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.906187 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.908269 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:34 crc kubenswrapper[4578]: E1003 12:51:34.908450 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.919607 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.942295 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.965375 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.981206 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:34 crc kubenswrapper[4578]: I1003 12:51:34.998264 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.019666 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/12c73964-5ff0-4b3a-8ec6-774035e92340-serviceca\") pod \"node-ca-rh9f2\" (UID: \"12c73964-5ff0-4b3a-8ec6-774035e92340\") " pod="openshift-image-registry/node-ca-rh9f2" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.020041 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf7m6\" (UniqueName: \"kubernetes.io/projected/12c73964-5ff0-4b3a-8ec6-774035e92340-kube-api-access-zf7m6\") pod \"node-ca-rh9f2\" (UID: \"12c73964-5ff0-4b3a-8ec6-774035e92340\") " pod="openshift-image-registry/node-ca-rh9f2" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.020060 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/12c73964-5ff0-4b3a-8ec6-774035e92340-host\") pod \"node-ca-rh9f2\" (UID: \"12c73964-5ff0-4b3a-8ec6-774035e92340\") " pod="openshift-image-registry/node-ca-rh9f2" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.021054 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.036748 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.052849 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.065845 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.076700 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.086853 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.101003 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.116735 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.121781 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/12c73964-5ff0-4b3a-8ec6-774035e92340-host\") pod \"node-ca-rh9f2\" (UID: \"12c73964-5ff0-4b3a-8ec6-774035e92340\") " pod="openshift-image-registry/node-ca-rh9f2" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.121828 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/12c73964-5ff0-4b3a-8ec6-774035e92340-serviceca\") pod \"node-ca-rh9f2\" (UID: \"12c73964-5ff0-4b3a-8ec6-774035e92340\") " pod="openshift-image-registry/node-ca-rh9f2" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.121862 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zf7m6\" (UniqueName: \"kubernetes.io/projected/12c73964-5ff0-4b3a-8ec6-774035e92340-kube-api-access-zf7m6\") pod \"node-ca-rh9f2\" (UID: \"12c73964-5ff0-4b3a-8ec6-774035e92340\") " pod="openshift-image-registry/node-ca-rh9f2" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.122341 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/12c73964-5ff0-4b3a-8ec6-774035e92340-host\") pod \"node-ca-rh9f2\" (UID: \"12c73964-5ff0-4b3a-8ec6-774035e92340\") " pod="openshift-image-registry/node-ca-rh9f2" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.126147 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/12c73964-5ff0-4b3a-8ec6-774035e92340-serviceca\") pod \"node-ca-rh9f2\" (UID: \"12c73964-5ff0-4b3a-8ec6-774035e92340\") " pod="openshift-image-registry/node-ca-rh9f2" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.130860 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerStarted","Data":"4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81"} Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.132493 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" event={"ID":"5403769c-f73e-4921-a402-3d650774c2e7","Type":"ContainerStarted","Data":"0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9"} Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.134826 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.137851 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d"} Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.148390 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.148426 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf7m6\" (UniqueName: \"kubernetes.io/projected/12c73964-5ff0-4b3a-8ec6-774035e92340-kube-api-access-zf7m6\") pod \"node-ca-rh9f2\" (UID: \"12c73964-5ff0-4b3a-8ec6-774035e92340\") " pod="openshift-image-registry/node-ca-rh9f2" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.164450 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.186308 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.198359 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.209285 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.218382 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.230437 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.248207 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.257715 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.268526 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.281436 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.302386 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.319999 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.333028 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.342664 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.381239 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.381447 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-rh9f2" Oct 03 12:51:35 crc kubenswrapper[4578]: W1003 12:51:35.416455 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12c73964_5ff0_4b3a_8ec6_774035e92340.slice/crio-0298f743634dc6d46729fa98574e8321ca705dca74008813d56b2061ff980c2f WatchSource:0}: Error finding container 0298f743634dc6d46729fa98574e8321ca705dca74008813d56b2061ff980c2f: Status 404 returned error can't find the container with id 0298f743634dc6d46729fa98574e8321ca705dca74008813d56b2061ff980c2f Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.430069 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.455966 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.473879 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.490068 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.526513 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:51:35 crc kubenswrapper[4578]: E1003 12:51:35.526827 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:51:39.526800862 +0000 UTC m=+35.325273086 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.534017 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.551249 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.572936 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.593786 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.620544 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.627904 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.627969 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.628002 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.628037 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:35 crc kubenswrapper[4578]: E1003 12:51:35.628184 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 12:51:35 crc kubenswrapper[4578]: E1003 12:51:35.628208 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 12:51:35 crc kubenswrapper[4578]: E1003 12:51:35.628238 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 12:51:35 crc kubenswrapper[4578]: E1003 12:51:35.628250 4578 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:35 crc kubenswrapper[4578]: E1003 12:51:35.628207 4578 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 12:51:35 crc kubenswrapper[4578]: E1003 12:51:35.628326 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:39.628308668 +0000 UTC m=+35.426780852 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:35 crc kubenswrapper[4578]: E1003 12:51:35.628361 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:39.628347088 +0000 UTC m=+35.426819272 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 12:51:35 crc kubenswrapper[4578]: E1003 12:51:35.628222 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 12:51:35 crc kubenswrapper[4578]: E1003 12:51:35.628411 4578 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:35 crc kubenswrapper[4578]: E1003 12:51:35.628444 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:39.628436001 +0000 UTC m=+35.426908285 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:35 crc kubenswrapper[4578]: E1003 12:51:35.628183 4578 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 12:51:35 crc kubenswrapper[4578]: E1003 12:51:35.628508 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:39.628499514 +0000 UTC m=+35.426971698 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.908650 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:35 crc kubenswrapper[4578]: E1003 12:51:35.908771 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:51:35 crc kubenswrapper[4578]: I1003 12:51:35.908819 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:35 crc kubenswrapper[4578]: E1003 12:51:35.908856 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.150718 4578 generic.go:334] "Generic (PLEG): container finished" podID="5403769c-f73e-4921-a402-3d650774c2e7" containerID="0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9" exitCode=0 Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.150815 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" event={"ID":"5403769c-f73e-4921-a402-3d650774c2e7","Type":"ContainerDied","Data":"0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9"} Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.155830 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-rh9f2" event={"ID":"12c73964-5ff0-4b3a-8ec6-774035e92340","Type":"ContainerStarted","Data":"e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d"} Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.155926 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-rh9f2" event={"ID":"12c73964-5ff0-4b3a-8ec6-774035e92340","Type":"ContainerStarted","Data":"0298f743634dc6d46729fa98574e8321ca705dca74008813d56b2061ff980c2f"} Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.177006 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.193098 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.206508 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.222736 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.238950 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.260617 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.274820 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.304661 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.322466 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.336875 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.362761 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.402868 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.438077 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.467749 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.498946 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.514989 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.534059 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.545274 4578 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.547345 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.547406 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.547421 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.547537 4578 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.553969 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.554952 4578 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.555307 4578 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.556242 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.556265 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.556274 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.556286 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.556296 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:36Z","lastTransitionTime":"2025-10-03T12:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.571319 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: E1003 12:51:36.576085 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.579385 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.579413 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.579421 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.579434 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.579444 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:36Z","lastTransitionTime":"2025-10-03T12:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:36 crc kubenswrapper[4578]: E1003 12:51:36.592038 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.592729 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.596774 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.596806 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.596833 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.596848 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.596860 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:36Z","lastTransitionTime":"2025-10-03T12:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.613295 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: E1003 12:51:36.615818 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.621073 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.621108 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.621118 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.621135 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.621146 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:36Z","lastTransitionTime":"2025-10-03T12:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.626513 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: E1003 12:51:36.632579 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.635936 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.635976 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.635986 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.636000 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.636009 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:36Z","lastTransitionTime":"2025-10-03T12:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.638178 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.650115 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: E1003 12:51:36.650520 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: E1003 12:51:36.650676 4578 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.652416 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.652449 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.652459 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.652472 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.652482 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:36Z","lastTransitionTime":"2025-10-03T12:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.664521 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.689169 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:36Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.754776 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.754819 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.754829 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.754845 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.754857 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:36Z","lastTransitionTime":"2025-10-03T12:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.856782 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.856821 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.856832 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.856847 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.856857 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:36Z","lastTransitionTime":"2025-10-03T12:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.908500 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:36 crc kubenswrapper[4578]: E1003 12:51:36.908627 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.959152 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.959189 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.959199 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.959213 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:36 crc kubenswrapper[4578]: I1003 12:51:36.959226 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:36Z","lastTransitionTime":"2025-10-03T12:51:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.061609 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.061677 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.061690 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.061707 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.061719 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:37Z","lastTransitionTime":"2025-10-03T12:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.162930 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerStarted","Data":"377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40"} Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.163375 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.163429 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.163442 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.163455 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.163467 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:37Z","lastTransitionTime":"2025-10-03T12:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.164899 4578 generic.go:334] "Generic (PLEG): container finished" podID="5403769c-f73e-4921-a402-3d650774c2e7" containerID="3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc" exitCode=0 Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.164951 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" event={"ID":"5403769c-f73e-4921-a402-3d650774c2e7","Type":"ContainerDied","Data":"3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc"} Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.184315 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:37Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.211073 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:37Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.228330 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:37Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.243774 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:37Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.259098 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:37Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.267123 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.267152 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.267161 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.267174 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.267183 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:37Z","lastTransitionTime":"2025-10-03T12:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.274262 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:37Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.287873 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:37Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.299416 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:37Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.316299 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:37Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.330891 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:37Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.345100 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:37Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.362358 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:37Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.369293 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.369472 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.369594 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.369717 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.369818 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:37Z","lastTransitionTime":"2025-10-03T12:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.378121 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:37Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.473333 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.473388 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.473402 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.473422 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.473439 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:37Z","lastTransitionTime":"2025-10-03T12:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.581073 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.581130 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.581184 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.581203 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.581215 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:37Z","lastTransitionTime":"2025-10-03T12:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.684446 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.684512 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.684530 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.684556 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.684574 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:37Z","lastTransitionTime":"2025-10-03T12:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.787288 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.787345 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.787353 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.787367 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.787430 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:37Z","lastTransitionTime":"2025-10-03T12:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.889954 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.890013 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.890025 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.890038 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.890049 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:37Z","lastTransitionTime":"2025-10-03T12:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.909092 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.909106 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:37 crc kubenswrapper[4578]: E1003 12:51:37.909342 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:51:37 crc kubenswrapper[4578]: E1003 12:51:37.909492 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.993934 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.994213 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.994332 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.994425 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:37 crc kubenswrapper[4578]: I1003 12:51:37.994539 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:37Z","lastTransitionTime":"2025-10-03T12:51:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.098598 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.098939 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.099040 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.099176 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.099277 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:38Z","lastTransitionTime":"2025-10-03T12:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.172922 4578 generic.go:334] "Generic (PLEG): container finished" podID="5403769c-f73e-4921-a402-3d650774c2e7" containerID="625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c" exitCode=0 Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.173171 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" event={"ID":"5403769c-f73e-4921-a402-3d650774c2e7","Type":"ContainerDied","Data":"625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c"} Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.198828 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:38Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.202037 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.202110 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.202131 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.202159 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.202178 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:38Z","lastTransitionTime":"2025-10-03T12:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.217795 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:38Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.239934 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:38Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.263819 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:38Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.286431 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:38Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.300907 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:38Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.305371 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.305470 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.305484 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.305502 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.305514 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:38Z","lastTransitionTime":"2025-10-03T12:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.318357 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:38Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.335836 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:38Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.354397 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:38Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.369883 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:38Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.383505 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:38Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.397746 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:38Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.408473 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.408507 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.408516 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.408534 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.408544 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:38Z","lastTransitionTime":"2025-10-03T12:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.417886 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:38Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.510921 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.511156 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.511241 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.511311 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.511391 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:38Z","lastTransitionTime":"2025-10-03T12:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.613423 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.614075 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.614088 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.614105 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.614128 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:38Z","lastTransitionTime":"2025-10-03T12:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.717010 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.717050 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.717060 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.717073 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.717082 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:38Z","lastTransitionTime":"2025-10-03T12:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.782767 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.783915 4578 scope.go:117] "RemoveContainer" containerID="9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1" Oct 03 12:51:38 crc kubenswrapper[4578]: E1003 12:51:38.784146 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.819985 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.820035 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.820047 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.820062 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.820074 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:38Z","lastTransitionTime":"2025-10-03T12:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.908273 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:38 crc kubenswrapper[4578]: E1003 12:51:38.908401 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.922153 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.922188 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.922199 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.922210 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:38 crc kubenswrapper[4578]: I1003 12:51:38.922223 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:38Z","lastTransitionTime":"2025-10-03T12:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.024671 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.024719 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.024731 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.024746 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.024760 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:39Z","lastTransitionTime":"2025-10-03T12:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.127389 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.127431 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.127443 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.127459 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.127471 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:39Z","lastTransitionTime":"2025-10-03T12:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.179985 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerStarted","Data":"ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6"} Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.180141 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.183559 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" event={"ID":"5403769c-f73e-4921-a402-3d650774c2e7","Type":"ContainerStarted","Data":"e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e"} Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.197038 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.213252 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.216114 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.230019 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.230066 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.230083 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.230099 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.230109 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:39Z","lastTransitionTime":"2025-10-03T12:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.242000 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.255844 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.271075 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.283886 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.296768 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.310468 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.324567 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.333664 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.333699 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.333707 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.333721 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.333732 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:39Z","lastTransitionTime":"2025-10-03T12:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.336071 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.347033 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.358852 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.370754 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.383851 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.394830 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.403980 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.416011 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.426853 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.436585 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.436761 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.436842 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.436928 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.437000 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:39Z","lastTransitionTime":"2025-10-03T12:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.438324 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.451313 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.464740 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.479645 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.493089 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.507930 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.521278 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.539491 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.539517 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.539525 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.539537 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.539545 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:39Z","lastTransitionTime":"2025-10-03T12:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.541333 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:39Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.568883 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:51:39 crc kubenswrapper[4578]: E1003 12:51:39.569132 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:51:47.569106731 +0000 UTC m=+43.367578915 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.642470 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.642525 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.642537 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.642558 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.642569 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:39Z","lastTransitionTime":"2025-10-03T12:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.670492 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.670566 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.670598 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.670627 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:39 crc kubenswrapper[4578]: E1003 12:51:39.670718 4578 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 12:51:39 crc kubenswrapper[4578]: E1003 12:51:39.670743 4578 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 12:51:39 crc kubenswrapper[4578]: E1003 12:51:39.670791 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:47.670774851 +0000 UTC m=+43.469247035 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 12:51:39 crc kubenswrapper[4578]: E1003 12:51:39.670838 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:47.670819592 +0000 UTC m=+43.469291856 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 12:51:39 crc kubenswrapper[4578]: E1003 12:51:39.670835 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 12:51:39 crc kubenswrapper[4578]: E1003 12:51:39.670870 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 12:51:39 crc kubenswrapper[4578]: E1003 12:51:39.670883 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 12:51:39 crc kubenswrapper[4578]: E1003 12:51:39.670887 4578 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:39 crc kubenswrapper[4578]: E1003 12:51:39.670902 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 12:51:39 crc kubenswrapper[4578]: E1003 12:51:39.670914 4578 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:39 crc kubenswrapper[4578]: E1003 12:51:39.670970 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:47.670950097 +0000 UTC m=+43.469422321 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:39 crc kubenswrapper[4578]: E1003 12:51:39.670998 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:47.670987187 +0000 UTC m=+43.469459471 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.744327 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.744370 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.744379 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.744394 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.744404 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:39Z","lastTransitionTime":"2025-10-03T12:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.848119 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.848151 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.848159 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.848174 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.848183 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:39Z","lastTransitionTime":"2025-10-03T12:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.908491 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.908571 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:39 crc kubenswrapper[4578]: E1003 12:51:39.908718 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:51:39 crc kubenswrapper[4578]: E1003 12:51:39.908873 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.951758 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.951822 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.951835 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.951853 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:39 crc kubenswrapper[4578]: I1003 12:51:39.951866 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:39Z","lastTransitionTime":"2025-10-03T12:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.054979 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.055044 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.055064 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.055092 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.055114 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:40Z","lastTransitionTime":"2025-10-03T12:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.157887 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.157934 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.157948 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.157966 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.157980 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:40Z","lastTransitionTime":"2025-10-03T12:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.186962 4578 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.187527 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.210286 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.224554 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:40Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.238851 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:40Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.249010 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:40Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.260713 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.260758 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.260778 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.260801 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.260815 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:40Z","lastTransitionTime":"2025-10-03T12:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.264695 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:40Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.276907 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:40Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.287414 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:40Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.301182 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:40Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.316250 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:40Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.330664 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:40Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.342790 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:40Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.355010 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:40Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.363183 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.363403 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.363497 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.363609 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.363714 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:40Z","lastTransitionTime":"2025-10-03T12:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.372496 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:40Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.392493 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:40Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.466652 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.466701 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.466715 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.466734 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.466747 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:40Z","lastTransitionTime":"2025-10-03T12:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.569389 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.569434 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.569447 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.569462 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.569472 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:40Z","lastTransitionTime":"2025-10-03T12:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.671850 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.671896 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.671908 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.671924 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.671935 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:40Z","lastTransitionTime":"2025-10-03T12:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.775249 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.775297 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.775314 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.775340 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.775357 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:40Z","lastTransitionTime":"2025-10-03T12:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.879016 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.879396 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.879407 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.879422 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.879433 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:40Z","lastTransitionTime":"2025-10-03T12:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.908697 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:40 crc kubenswrapper[4578]: E1003 12:51:40.908840 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.991215 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.991259 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.991272 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.991292 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:40 crc kubenswrapper[4578]: I1003 12:51:40.991307 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:40Z","lastTransitionTime":"2025-10-03T12:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.094187 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.094280 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.094292 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.094308 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.094319 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:41Z","lastTransitionTime":"2025-10-03T12:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.193906 4578 generic.go:334] "Generic (PLEG): container finished" podID="5403769c-f73e-4921-a402-3d650774c2e7" containerID="e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e" exitCode=0 Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.193999 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" event={"ID":"5403769c-f73e-4921-a402-3d650774c2e7","Type":"ContainerDied","Data":"e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e"} Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.194483 4578 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.201321 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.201350 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.201358 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.201372 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.201381 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:41Z","lastTransitionTime":"2025-10-03T12:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.208921 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:41Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.220272 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:41Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.244951 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:41Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.260439 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:41Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.272714 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:41Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.283807 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:41Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.301681 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:41Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.305161 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.305195 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.305209 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.305223 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.305231 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:41Z","lastTransitionTime":"2025-10-03T12:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.314934 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:41Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.325821 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:41Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.335559 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:41Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.350658 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:41Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.362279 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:41Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.374066 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:41Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.407995 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.408037 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.408050 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.408067 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.408077 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:41Z","lastTransitionTime":"2025-10-03T12:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.510274 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.510314 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.510323 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.510337 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.510346 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:41Z","lastTransitionTime":"2025-10-03T12:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.613446 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.613483 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.613491 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.613505 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.613513 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:41Z","lastTransitionTime":"2025-10-03T12:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.716675 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.716752 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.716772 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.716800 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.716819 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:41Z","lastTransitionTime":"2025-10-03T12:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.819134 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.819180 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.819192 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.819207 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.819219 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:41Z","lastTransitionTime":"2025-10-03T12:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.908559 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.908593 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:41 crc kubenswrapper[4578]: E1003 12:51:41.908742 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:51:41 crc kubenswrapper[4578]: E1003 12:51:41.908912 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.922270 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.922309 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.922322 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.922353 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:41 crc kubenswrapper[4578]: I1003 12:51:41.922365 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:41Z","lastTransitionTime":"2025-10-03T12:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.025670 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.025732 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.025744 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.025764 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.025776 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:42Z","lastTransitionTime":"2025-10-03T12:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.128251 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.128296 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.128306 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.128321 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.128334 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:42Z","lastTransitionTime":"2025-10-03T12:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.201379 4578 generic.go:334] "Generic (PLEG): container finished" podID="5403769c-f73e-4921-a402-3d650774c2e7" containerID="f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6" exitCode=0 Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.201414 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" event={"ID":"5403769c-f73e-4921-a402-3d650774c2e7","Type":"ContainerDied","Data":"f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6"} Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.201542 4578 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.231078 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.231131 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.231148 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.231172 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.231188 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:42Z","lastTransitionTime":"2025-10-03T12:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.237263 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:42Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.262280 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:42Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.283475 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:42Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.299771 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:42Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.315640 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:42Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.331489 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:42Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.337840 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.337885 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.337899 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.337921 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.337935 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:42Z","lastTransitionTime":"2025-10-03T12:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.347672 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:42Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.363353 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:42Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.379679 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:42Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.435163 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:42Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.440874 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.440926 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.440943 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.440972 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.440992 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:42Z","lastTransitionTime":"2025-10-03T12:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.450093 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:42Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.465675 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:42Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.485432 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:42Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.543864 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.543934 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.543949 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.543978 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.543995 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:42Z","lastTransitionTime":"2025-10-03T12:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.646729 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.646770 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.646783 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.646803 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.646819 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:42Z","lastTransitionTime":"2025-10-03T12:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.749374 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.749428 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.749442 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.749461 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.749473 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:42Z","lastTransitionTime":"2025-10-03T12:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.853007 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.853063 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.853080 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.853102 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.853117 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:42Z","lastTransitionTime":"2025-10-03T12:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.908770 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:42 crc kubenswrapper[4578]: E1003 12:51:42.908995 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.955444 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.955484 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.955493 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.955511 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:42 crc kubenswrapper[4578]: I1003 12:51:42.955521 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:42Z","lastTransitionTime":"2025-10-03T12:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.058928 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.059599 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.059835 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.059881 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.059907 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:43Z","lastTransitionTime":"2025-10-03T12:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.161901 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.161935 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.161947 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.161962 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.161973 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:43Z","lastTransitionTime":"2025-10-03T12:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.209752 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" event={"ID":"5403769c-f73e-4921-a402-3d650774c2e7","Type":"ContainerStarted","Data":"586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2"} Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.225903 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:43Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.237380 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:43Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.246581 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:43Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.262616 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:43Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.264283 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.264306 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.264316 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.264331 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.264339 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:43Z","lastTransitionTime":"2025-10-03T12:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.275006 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:43Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.287368 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:43Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.299610 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:43Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.311408 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:43Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.324086 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:43Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.338192 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:43Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.352530 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:43Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.365176 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:43Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.366355 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.366402 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.366413 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.366430 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.366439 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:43Z","lastTransitionTime":"2025-10-03T12:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.383767 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:43Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.469269 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.469318 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.469330 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.469349 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.469364 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:43Z","lastTransitionTime":"2025-10-03T12:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.572219 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.572274 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.572284 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.572300 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.572309 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:43Z","lastTransitionTime":"2025-10-03T12:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.675015 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.675082 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.675095 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.675119 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.675134 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:43Z","lastTransitionTime":"2025-10-03T12:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.778152 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.778215 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.778229 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.778254 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.778272 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:43Z","lastTransitionTime":"2025-10-03T12:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.880450 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.880480 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.880488 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.880502 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.880510 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:43Z","lastTransitionTime":"2025-10-03T12:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.908210 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.908277 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:43 crc kubenswrapper[4578]: E1003 12:51:43.908411 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:51:43 crc kubenswrapper[4578]: E1003 12:51:43.908578 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.983467 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.983521 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.983533 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.983556 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:43 crc kubenswrapper[4578]: I1003 12:51:43.983571 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:43Z","lastTransitionTime":"2025-10-03T12:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.087066 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.087110 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.087123 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.087144 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.087160 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:44Z","lastTransitionTime":"2025-10-03T12:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.189853 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.189888 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.189898 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.189913 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.189925 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:44Z","lastTransitionTime":"2025-10-03T12:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.214815 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8qvh9_ebaa96c5-be7c-4f05-819d-f37e1ceddac8/ovnkube-controller/0.log" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.218203 4578 generic.go:334] "Generic (PLEG): container finished" podID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerID="ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6" exitCode=1 Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.218245 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerDied","Data":"ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6"} Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.219003 4578 scope.go:117] "RemoveContainer" containerID="ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.232863 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9"] Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.233261 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.234601 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.234885 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.235195 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.252247 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.263140 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.275012 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.287376 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.291586 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.291647 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.291659 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.291674 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.291685 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:44Z","lastTransitionTime":"2025-10-03T12:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.306219 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"message\\\":\\\"work-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1003 12:51:43.258428 5763 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258471 5763 factory.go:656] Stopping watch factory\\\\nI1003 12:51:43.258513 5763 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258603 5763 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258755 5763 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259087 5763 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259307 5763 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259461 5763 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 12:51:43.259545 5763 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.314541 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e95e1c26-64e1-467b-bf8d-aa066847a95a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-xc4j9\" (UID: \"e95e1c26-64e1-467b-bf8d-aa066847a95a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.314589 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e95e1c26-64e1-467b-bf8d-aa066847a95a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-xc4j9\" (UID: \"e95e1c26-64e1-467b-bf8d-aa066847a95a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.314657 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr4tx\" (UniqueName: \"kubernetes.io/projected/e95e1c26-64e1-467b-bf8d-aa066847a95a-kube-api-access-sr4tx\") pod \"ovnkube-control-plane-749d76644c-xc4j9\" (UID: \"e95e1c26-64e1-467b-bf8d-aa066847a95a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.314682 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e95e1c26-64e1-467b-bf8d-aa066847a95a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-xc4j9\" (UID: \"e95e1c26-64e1-467b-bf8d-aa066847a95a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.319529 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.332747 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.343584 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.353180 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.364181 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.376440 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.387870 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.393400 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.393433 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.393446 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.393463 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.393474 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:44Z","lastTransitionTime":"2025-10-03T12:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.399423 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.409025 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.417118 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e95e1c26-64e1-467b-bf8d-aa066847a95a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-xc4j9\" (UID: \"e95e1c26-64e1-467b-bf8d-aa066847a95a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.417160 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e95e1c26-64e1-467b-bf8d-aa066847a95a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-xc4j9\" (UID: \"e95e1c26-64e1-467b-bf8d-aa066847a95a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.417205 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr4tx\" (UniqueName: \"kubernetes.io/projected/e95e1c26-64e1-467b-bf8d-aa066847a95a-kube-api-access-sr4tx\") pod \"ovnkube-control-plane-749d76644c-xc4j9\" (UID: \"e95e1c26-64e1-467b-bf8d-aa066847a95a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.417234 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e95e1c26-64e1-467b-bf8d-aa066847a95a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-xc4j9\" (UID: \"e95e1c26-64e1-467b-bf8d-aa066847a95a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.417783 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e95e1c26-64e1-467b-bf8d-aa066847a95a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-xc4j9\" (UID: \"e95e1c26-64e1-467b-bf8d-aa066847a95a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.418229 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e95e1c26-64e1-467b-bf8d-aa066847a95a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-xc4j9\" (UID: \"e95e1c26-64e1-467b-bf8d-aa066847a95a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.420433 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.427412 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e95e1c26-64e1-467b-bf8d-aa066847a95a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-xc4j9\" (UID: \"e95e1c26-64e1-467b-bf8d-aa066847a95a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.430622 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.433535 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr4tx\" (UniqueName: \"kubernetes.io/projected/e95e1c26-64e1-467b-bf8d-aa066847a95a-kube-api-access-sr4tx\") pod \"ovnkube-control-plane-749d76644c-xc4j9\" (UID: \"e95e1c26-64e1-467b-bf8d-aa066847a95a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.443137 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.455457 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.470737 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.485681 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.495336 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.495388 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.495398 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.495413 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.495422 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:44Z","lastTransitionTime":"2025-10-03T12:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.499979 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.511979 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.531225 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"message\\\":\\\"work-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1003 12:51:43.258428 5763 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258471 5763 factory.go:656] Stopping watch factory\\\\nI1003 12:51:43.258513 5763 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258603 5763 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258755 5763 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259087 5763 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259307 5763 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259461 5763 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 12:51:43.259545 5763 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.545129 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.550264 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.561386 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.572027 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.600733 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.600980 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.601037 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.601093 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.601143 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:44Z","lastTransitionTime":"2025-10-03T12:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.703799 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.703838 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.703849 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.703865 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.703876 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:44Z","lastTransitionTime":"2025-10-03T12:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.806127 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.806163 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.806171 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.806202 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.806212 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:44Z","lastTransitionTime":"2025-10-03T12:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.910728 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:44 crc kubenswrapper[4578]: E1003 12:51:44.910834 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.911141 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.911158 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.911166 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.911177 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.911186 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:44Z","lastTransitionTime":"2025-10-03T12:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.925840 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.940859 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.951763 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.963823 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.979126 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:44 crc kubenswrapper[4578]: I1003 12:51:44.989076 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.000942 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.012968 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.012993 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.013002 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.013029 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.013038 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:45Z","lastTransitionTime":"2025-10-03T12:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.013869 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.025905 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.041905 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"message\\\":\\\"work-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1003 12:51:43.258428 5763 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258471 5763 factory.go:656] Stopping watch factory\\\\nI1003 12:51:43.258513 5763 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258603 5763 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258755 5763 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259087 5763 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259307 5763 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259461 5763 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 12:51:43.259545 5763 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.053432 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.065358 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.080039 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.093377 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.118887 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.118943 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.118954 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.118973 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.118986 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:45Z","lastTransitionTime":"2025-10-03T12:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.227795 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.227957 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.227973 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.227994 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.228008 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:45Z","lastTransitionTime":"2025-10-03T12:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.229383 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8qvh9_ebaa96c5-be7c-4f05-819d-f37e1ceddac8/ovnkube-controller/0.log" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.232219 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerStarted","Data":"2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5"} Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.232352 4578 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.233831 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" event={"ID":"e95e1c26-64e1-467b-bf8d-aa066847a95a","Type":"ContainerStarted","Data":"38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177"} Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.233856 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" event={"ID":"e95e1c26-64e1-467b-bf8d-aa066847a95a","Type":"ContainerStarted","Data":"385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1"} Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.233866 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" event={"ID":"e95e1c26-64e1-467b-bf8d-aa066847a95a","Type":"ContainerStarted","Data":"26f4dccca09d6d5d2ec2fd8cd464ee2de83bbb4a627e6abfcb89efeb48b96d00"} Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.251969 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.270420 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.293369 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.308785 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-q8986"] Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.309300 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:51:45 crc kubenswrapper[4578]: E1003 12:51:45.309364 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.312212 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"message\\\":\\\"work-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1003 12:51:43.258428 5763 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258471 5763 factory.go:656] Stopping watch factory\\\\nI1003 12:51:43.258513 5763 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258603 5763 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258755 5763 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259087 5763 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259307 5763 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259461 5763 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 12:51:43.259545 5763 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.324772 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.330362 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.330420 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.330450 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.330468 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.330482 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:45Z","lastTransitionTime":"2025-10-03T12:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.336347 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.347245 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.360623 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.375019 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.385793 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.396985 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.411515 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.425850 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x9d9\" (UniqueName: \"kubernetes.io/projected/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-kube-api-access-2x9d9\") pod \"network-metrics-daemon-q8986\" (UID: \"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\") " pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.425963 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs\") pod \"network-metrics-daemon-q8986\" (UID: \"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\") " pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.435744 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.437597 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.437990 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.438005 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.438026 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.438039 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:45Z","lastTransitionTime":"2025-10-03T12:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.455159 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.470912 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.491991 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.506155 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.523940 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.526610 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs\") pod \"network-metrics-daemon-q8986\" (UID: \"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\") " pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.526673 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x9d9\" (UniqueName: \"kubernetes.io/projected/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-kube-api-access-2x9d9\") pod \"network-metrics-daemon-q8986\" (UID: \"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\") " pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:51:45 crc kubenswrapper[4578]: E1003 12:51:45.526818 4578 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 12:51:45 crc kubenswrapper[4578]: E1003 12:51:45.526920 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs podName:2971c7bb-e0f4-44a8-8f52-9ea8ba199772 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:46.026890402 +0000 UTC m=+41.825362586 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs") pod "network-metrics-daemon-q8986" (UID: "2971c7bb-e0f4-44a8-8f52-9ea8ba199772") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.541029 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.541072 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.541083 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.541100 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.541115 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:45Z","lastTransitionTime":"2025-10-03T12:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.547107 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"message\\\":\\\"work-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1003 12:51:43.258428 5763 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258471 5763 factory.go:656] Stopping watch factory\\\\nI1003 12:51:43.258513 5763 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258603 5763 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258755 5763 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259087 5763 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259307 5763 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259461 5763 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 12:51:43.259545 5763 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.562122 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x9d9\" (UniqueName: \"kubernetes.io/projected/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-kube-api-access-2x9d9\") pod \"network-metrics-daemon-q8986\" (UID: \"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\") " pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.574068 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.595667 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.610478 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.625318 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.637056 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.642589 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.642699 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.642710 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.642725 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.642733 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:45Z","lastTransitionTime":"2025-10-03T12:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.652205 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.664105 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.675484 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.695142 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.708762 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.745484 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.745533 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.745549 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.745572 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.745608 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:45Z","lastTransitionTime":"2025-10-03T12:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.847949 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.847986 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.848000 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.848016 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.848027 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:45Z","lastTransitionTime":"2025-10-03T12:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.908565 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.908581 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:45 crc kubenswrapper[4578]: E1003 12:51:45.908758 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:51:45 crc kubenswrapper[4578]: E1003 12:51:45.908855 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.950598 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.950648 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.950657 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.950670 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:45 crc kubenswrapper[4578]: I1003 12:51:45.950679 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:45Z","lastTransitionTime":"2025-10-03T12:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.032575 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs\") pod \"network-metrics-daemon-q8986\" (UID: \"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\") " pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:51:46 crc kubenswrapper[4578]: E1003 12:51:46.032755 4578 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 12:51:46 crc kubenswrapper[4578]: E1003 12:51:46.032826 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs podName:2971c7bb-e0f4-44a8-8f52-9ea8ba199772 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:47.032807506 +0000 UTC m=+42.831279690 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs") pod "network-metrics-daemon-q8986" (UID: "2971c7bb-e0f4-44a8-8f52-9ea8ba199772") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.053305 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.053348 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.053358 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.053375 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.053413 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:46Z","lastTransitionTime":"2025-10-03T12:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.155311 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.155354 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.155365 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.155399 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.155412 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:46Z","lastTransitionTime":"2025-10-03T12:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.240287 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8qvh9_ebaa96c5-be7c-4f05-819d-f37e1ceddac8/ovnkube-controller/1.log" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.240947 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8qvh9_ebaa96c5-be7c-4f05-819d-f37e1ceddac8/ovnkube-controller/0.log" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.244103 4578 generic.go:334] "Generic (PLEG): container finished" podID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerID="2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5" exitCode=1 Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.244161 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerDied","Data":"2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5"} Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.244239 4578 scope.go:117] "RemoveContainer" containerID="ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.245165 4578 scope.go:117] "RemoveContainer" containerID="2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5" Oct 03 12:51:46 crc kubenswrapper[4578]: E1003 12:51:46.245390 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.258144 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.258208 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.258250 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.258269 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.258279 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:46Z","lastTransitionTime":"2025-10-03T12:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.263565 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.278886 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.294226 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.306778 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.316854 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.333751 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"message\\\":\\\"work-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1003 12:51:43.258428 5763 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258471 5763 factory.go:656] Stopping watch factory\\\\nI1003 12:51:43.258513 5763 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258603 5763 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258755 5763 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259087 5763 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259307 5763 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259461 5763 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 12:51:43.259545 5763 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.344564 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.358680 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.360048 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.360104 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.360131 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.360159 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.360170 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:46Z","lastTransitionTime":"2025-10-03T12:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.372159 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.382616 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.392447 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.404600 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.414823 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.428154 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.446019 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.460442 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.461880 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.461923 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.461932 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.461953 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.461962 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:46Z","lastTransitionTime":"2025-10-03T12:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.472976 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.486840 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.506702 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"message\\\":\\\"work-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1003 12:51:43.258428 5763 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258471 5763 factory.go:656] Stopping watch factory\\\\nI1003 12:51:43.258513 5763 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258603 5763 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258755 5763 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259087 5763 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259307 5763 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259461 5763 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 12:51:43.259545 5763 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"vices.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:51:45.973041 5964 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}\\\\nI1003 12:51:45.976192 5964 services_controller.go:444] Built service openshift-kube-apiserver-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1003 12:51:45.976195 5964 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1003 12:51:45.976169 5964 model_client.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.522203 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.534160 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.544163 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.552306 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.562698 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.564120 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.564186 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.564201 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.564216 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.564225 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:46Z","lastTransitionTime":"2025-10-03T12:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.573698 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.585587 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.597076 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.607072 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.618426 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.631322 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.666309 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.666353 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.666368 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.666384 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.666394 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:46Z","lastTransitionTime":"2025-10-03T12:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.693026 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.693072 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.693082 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.693094 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.693102 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:46Z","lastTransitionTime":"2025-10-03T12:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:46 crc kubenswrapper[4578]: E1003 12:51:46.705337 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.711524 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.711563 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.711574 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.711590 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.711602 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:46Z","lastTransitionTime":"2025-10-03T12:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:46 crc kubenswrapper[4578]: E1003 12:51:46.727652 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.730911 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.730955 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.730964 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.730979 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.730989 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:46Z","lastTransitionTime":"2025-10-03T12:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:46 crc kubenswrapper[4578]: E1003 12:51:46.746835 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.750063 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.750105 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.750114 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.750128 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.750138 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:46Z","lastTransitionTime":"2025-10-03T12:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:46 crc kubenswrapper[4578]: E1003 12:51:46.763553 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.767293 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.767331 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.767343 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.767360 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.767372 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:46Z","lastTransitionTime":"2025-10-03T12:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:46 crc kubenswrapper[4578]: E1003 12:51:46.780683 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:46Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:46 crc kubenswrapper[4578]: E1003 12:51:46.780810 4578 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.782291 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.782329 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.782337 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.782351 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.782360 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:46Z","lastTransitionTime":"2025-10-03T12:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.884517 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.884551 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.884562 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.884576 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.884621 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:46Z","lastTransitionTime":"2025-10-03T12:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.909107 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:46 crc kubenswrapper[4578]: E1003 12:51:46.909332 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.909561 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:51:46 crc kubenswrapper[4578]: E1003 12:51:46.909749 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.987272 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.987329 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.987341 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.987357 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:46 crc kubenswrapper[4578]: I1003 12:51:46.987365 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:46Z","lastTransitionTime":"2025-10-03T12:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.042180 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs\") pod \"network-metrics-daemon-q8986\" (UID: \"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\") " pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:51:47 crc kubenswrapper[4578]: E1003 12:51:47.042327 4578 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 12:51:47 crc kubenswrapper[4578]: E1003 12:51:47.042398 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs podName:2971c7bb-e0f4-44a8-8f52-9ea8ba199772 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:49.042363323 +0000 UTC m=+44.840835497 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs") pod "network-metrics-daemon-q8986" (UID: "2971c7bb-e0f4-44a8-8f52-9ea8ba199772") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.089746 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.089790 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.089802 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.089818 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.089829 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:47Z","lastTransitionTime":"2025-10-03T12:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.193148 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.193195 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.193206 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.193221 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.193232 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:47Z","lastTransitionTime":"2025-10-03T12:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.248532 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8qvh9_ebaa96c5-be7c-4f05-819d-f37e1ceddac8/ovnkube-controller/1.log" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.296038 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.296079 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.296094 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.296113 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.296128 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:47Z","lastTransitionTime":"2025-10-03T12:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.398292 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.398321 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.398329 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.398342 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.398351 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:47Z","lastTransitionTime":"2025-10-03T12:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.500965 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.501007 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.501016 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.501029 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.501038 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:47Z","lastTransitionTime":"2025-10-03T12:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.605835 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.605895 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.605907 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.605927 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.605939 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:47Z","lastTransitionTime":"2025-10-03T12:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.648128 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:51:47 crc kubenswrapper[4578]: E1003 12:51:47.648325 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:52:03.648300025 +0000 UTC m=+59.446772209 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.707909 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.707952 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.707963 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.707976 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.707984 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:47Z","lastTransitionTime":"2025-10-03T12:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.749845 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.749905 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.749936 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.749979 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:47 crc kubenswrapper[4578]: E1003 12:51:47.750029 4578 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 12:51:47 crc kubenswrapper[4578]: E1003 12:51:47.750062 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 12:51:47 crc kubenswrapper[4578]: E1003 12:51:47.750086 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 12:51:47 crc kubenswrapper[4578]: E1003 12:51:47.750088 4578 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 12:51:47 crc kubenswrapper[4578]: E1003 12:51:47.750098 4578 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:47 crc kubenswrapper[4578]: E1003 12:51:47.750122 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 12:52:03.75010238 +0000 UTC m=+59.548574574 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 12:51:47 crc kubenswrapper[4578]: E1003 12:51:47.750141 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 12:52:03.750132201 +0000 UTC m=+59.548604385 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 12:51:47 crc kubenswrapper[4578]: E1003 12:51:47.750154 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 12:52:03.750148172 +0000 UTC m=+59.548620426 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:47 crc kubenswrapper[4578]: E1003 12:51:47.750202 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 12:51:47 crc kubenswrapper[4578]: E1003 12:51:47.750218 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 12:51:47 crc kubenswrapper[4578]: E1003 12:51:47.750231 4578 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:47 crc kubenswrapper[4578]: E1003 12:51:47.750267 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 12:52:03.750251615 +0000 UTC m=+59.548723839 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.810734 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.810790 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.810805 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.810823 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.810835 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:47Z","lastTransitionTime":"2025-10-03T12:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.908845 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.909111 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:47 crc kubenswrapper[4578]: E1003 12:51:47.909224 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:51:47 crc kubenswrapper[4578]: E1003 12:51:47.909278 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.914110 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.914165 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.914178 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.914195 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:47 crc kubenswrapper[4578]: I1003 12:51:47.914207 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:47Z","lastTransitionTime":"2025-10-03T12:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.017167 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.017226 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.017236 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.017254 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.017266 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:48Z","lastTransitionTime":"2025-10-03T12:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.120241 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.120311 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.120333 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.120362 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.120383 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:48Z","lastTransitionTime":"2025-10-03T12:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.222763 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.222821 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.222831 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.222844 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.222859 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:48Z","lastTransitionTime":"2025-10-03T12:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.329794 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.329863 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.329873 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.329887 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.329896 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:48Z","lastTransitionTime":"2025-10-03T12:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.432787 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.432831 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.432842 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.432856 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.432865 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:48Z","lastTransitionTime":"2025-10-03T12:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.535888 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.536267 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.536370 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.536453 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.536528 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:48Z","lastTransitionTime":"2025-10-03T12:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.639406 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.639443 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.639455 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.639472 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.639485 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:48Z","lastTransitionTime":"2025-10-03T12:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.742698 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.742770 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.742780 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.742792 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.743276 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:48Z","lastTransitionTime":"2025-10-03T12:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.845709 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.845740 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.845789 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.845805 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.845814 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:48Z","lastTransitionTime":"2025-10-03T12:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.909153 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.909222 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:51:48 crc kubenswrapper[4578]: E1003 12:51:48.909351 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:51:48 crc kubenswrapper[4578]: E1003 12:51:48.909490 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.948104 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.948279 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.948340 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.948445 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:48 crc kubenswrapper[4578]: I1003 12:51:48.948507 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:48Z","lastTransitionTime":"2025-10-03T12:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.051930 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.051996 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.052015 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.052039 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.052057 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:49Z","lastTransitionTime":"2025-10-03T12:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.064021 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs\") pod \"network-metrics-daemon-q8986\" (UID: \"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\") " pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:51:49 crc kubenswrapper[4578]: E1003 12:51:49.064276 4578 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 12:51:49 crc kubenswrapper[4578]: E1003 12:51:49.064346 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs podName:2971c7bb-e0f4-44a8-8f52-9ea8ba199772 nodeName:}" failed. No retries permitted until 2025-10-03 12:51:53.064325628 +0000 UTC m=+48.862797812 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs") pod "network-metrics-daemon-q8986" (UID: "2971c7bb-e0f4-44a8-8f52-9ea8ba199772") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.153969 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.154002 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.154011 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.154025 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.154035 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:49Z","lastTransitionTime":"2025-10-03T12:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.255802 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.255858 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.255869 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.255889 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.255904 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:49Z","lastTransitionTime":"2025-10-03T12:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.359104 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.359142 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.359152 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.359165 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.359175 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:49Z","lastTransitionTime":"2025-10-03T12:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.462489 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.462527 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.462535 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.462548 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.462557 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:49Z","lastTransitionTime":"2025-10-03T12:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.564498 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.564527 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.564534 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.564547 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.564554 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:49Z","lastTransitionTime":"2025-10-03T12:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.666822 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.666863 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.666872 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.666887 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.666898 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:49Z","lastTransitionTime":"2025-10-03T12:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.769826 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.769867 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.769876 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.769888 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.769896 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:49Z","lastTransitionTime":"2025-10-03T12:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.873496 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.873540 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.873552 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.873572 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.873584 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:49Z","lastTransitionTime":"2025-10-03T12:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.908554 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.908617 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:49 crc kubenswrapper[4578]: E1003 12:51:49.908801 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:51:49 crc kubenswrapper[4578]: E1003 12:51:49.908986 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.975948 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.975986 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.975997 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.976014 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:49 crc kubenswrapper[4578]: I1003 12:51:49.976026 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:49Z","lastTransitionTime":"2025-10-03T12:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.078615 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.078670 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.078680 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.078694 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.078703 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:50Z","lastTransitionTime":"2025-10-03T12:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.181457 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.181516 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.181568 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.181584 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.181594 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:50Z","lastTransitionTime":"2025-10-03T12:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.284240 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.284287 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.284297 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.284312 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.284323 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:50Z","lastTransitionTime":"2025-10-03T12:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.387815 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.387853 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.387863 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.387876 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.387885 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:50Z","lastTransitionTime":"2025-10-03T12:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.490250 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.490291 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.490304 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.490321 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.490333 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:50Z","lastTransitionTime":"2025-10-03T12:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.592252 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.592291 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.592301 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.592314 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.592323 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:50Z","lastTransitionTime":"2025-10-03T12:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.695088 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.695148 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.695166 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.695187 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.695203 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:50Z","lastTransitionTime":"2025-10-03T12:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.797548 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.797828 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.797936 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.798015 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.798128 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:50Z","lastTransitionTime":"2025-10-03T12:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.901510 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.901559 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.901573 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.901591 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.901608 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:50Z","lastTransitionTime":"2025-10-03T12:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.908602 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:51:50 crc kubenswrapper[4578]: E1003 12:51:50.908753 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:51:50 crc kubenswrapper[4578]: I1003 12:51:50.908917 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:50 crc kubenswrapper[4578]: E1003 12:51:50.909052 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.005627 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.005679 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.005687 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.005700 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.005709 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:51Z","lastTransitionTime":"2025-10-03T12:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.108618 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.108683 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.108694 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.108709 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.108721 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:51Z","lastTransitionTime":"2025-10-03T12:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.211650 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.212000 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.212116 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.212221 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.212304 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:51Z","lastTransitionTime":"2025-10-03T12:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.314476 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.314517 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.314532 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.314554 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.314568 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:51Z","lastTransitionTime":"2025-10-03T12:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.416821 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.417222 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.417235 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.417256 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.417278 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:51Z","lastTransitionTime":"2025-10-03T12:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.522375 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.522416 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.522425 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.522440 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.522451 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:51Z","lastTransitionTime":"2025-10-03T12:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.625365 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.625408 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.625422 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.625446 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.625459 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:51Z","lastTransitionTime":"2025-10-03T12:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.728240 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.728470 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.728530 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.728591 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.728662 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:51Z","lastTransitionTime":"2025-10-03T12:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.831692 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.832143 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.832317 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.832475 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.832653 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:51Z","lastTransitionTime":"2025-10-03T12:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.908762 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:51 crc kubenswrapper[4578]: E1003 12:51:51.909421 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.908788 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:51 crc kubenswrapper[4578]: E1003 12:51:51.910362 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.936432 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.936479 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.936491 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.936511 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:51 crc kubenswrapper[4578]: I1003 12:51:51.936524 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:51Z","lastTransitionTime":"2025-10-03T12:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.039375 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.039431 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.039445 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.039463 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.039475 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:52Z","lastTransitionTime":"2025-10-03T12:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.142625 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.142680 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.142689 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.142704 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.142713 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:52Z","lastTransitionTime":"2025-10-03T12:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.245155 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.245604 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.245711 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.245793 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.245868 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:52Z","lastTransitionTime":"2025-10-03T12:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.349011 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.349094 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.349122 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.349157 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.349181 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:52Z","lastTransitionTime":"2025-10-03T12:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.451433 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.451469 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.451477 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.451492 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.451501 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:52Z","lastTransitionTime":"2025-10-03T12:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.553820 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.553877 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.553893 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.553914 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.553929 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:52Z","lastTransitionTime":"2025-10-03T12:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.656797 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.656854 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.656865 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.656880 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.656890 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:52Z","lastTransitionTime":"2025-10-03T12:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.760538 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.760588 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.760600 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.760617 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.760656 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:52Z","lastTransitionTime":"2025-10-03T12:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.863198 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.863258 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.863270 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.863291 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.863305 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:52Z","lastTransitionTime":"2025-10-03T12:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.908075 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.908100 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:51:52 crc kubenswrapper[4578]: E1003 12:51:52.908257 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:51:52 crc kubenswrapper[4578]: E1003 12:51:52.908370 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.966163 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.966214 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.966225 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.966241 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:52 crc kubenswrapper[4578]: I1003 12:51:52.966253 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:52Z","lastTransitionTime":"2025-10-03T12:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.069117 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.069158 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.069170 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.069184 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.069193 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:53Z","lastTransitionTime":"2025-10-03T12:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.106852 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs\") pod \"network-metrics-daemon-q8986\" (UID: \"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\") " pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:51:53 crc kubenswrapper[4578]: E1003 12:51:53.107058 4578 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 12:51:53 crc kubenswrapper[4578]: E1003 12:51:53.107169 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs podName:2971c7bb-e0f4-44a8-8f52-9ea8ba199772 nodeName:}" failed. No retries permitted until 2025-10-03 12:52:01.107143111 +0000 UTC m=+56.905615295 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs") pod "network-metrics-daemon-q8986" (UID: "2971c7bb-e0f4-44a8-8f52-9ea8ba199772") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.172323 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.172376 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.172388 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.172408 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.172425 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:53Z","lastTransitionTime":"2025-10-03T12:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.277823 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.277923 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.278454 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.278490 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.278500 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:53Z","lastTransitionTime":"2025-10-03T12:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.381217 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.381489 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.381556 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.381649 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.381729 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:53Z","lastTransitionTime":"2025-10-03T12:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.484595 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.484641 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.484650 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.484663 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.484671 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:53Z","lastTransitionTime":"2025-10-03T12:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.587534 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.587574 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.587584 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.587600 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.587611 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:53Z","lastTransitionTime":"2025-10-03T12:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.690046 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.690086 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.690098 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.690117 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.690129 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:53Z","lastTransitionTime":"2025-10-03T12:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.793040 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.793083 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.793094 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.793151 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.793162 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:53Z","lastTransitionTime":"2025-10-03T12:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.895761 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.895800 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.895809 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.895823 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.895833 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:53Z","lastTransitionTime":"2025-10-03T12:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.909006 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.909060 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:53 crc kubenswrapper[4578]: E1003 12:51:53.909134 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:51:53 crc kubenswrapper[4578]: E1003 12:51:53.909407 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.909723 4578 scope.go:117] "RemoveContainer" containerID="9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.998165 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.998200 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.998209 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.998232 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:53 crc kubenswrapper[4578]: I1003 12:51:53.998244 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:53Z","lastTransitionTime":"2025-10-03T12:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.101679 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.101712 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.101720 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.101736 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.101747 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:54Z","lastTransitionTime":"2025-10-03T12:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.205189 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.205274 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.205289 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.205310 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.205327 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:54Z","lastTransitionTime":"2025-10-03T12:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.271661 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.273582 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"93b4650253599f9031287cc43642dff608baa3920e6dce60b433a7c642cece88"} Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.274420 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.290598 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.303590 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.308018 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.308043 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.308052 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.308066 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.308075 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:54Z","lastTransitionTime":"2025-10-03T12:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.318617 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.333316 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b4650253599f9031287cc43642dff608baa3920e6dce60b433a7c642cece88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.348476 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.362170 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.380290 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"message\\\":\\\"work-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1003 12:51:43.258428 5763 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258471 5763 factory.go:656] Stopping watch factory\\\\nI1003 12:51:43.258513 5763 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258603 5763 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258755 5763 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259087 5763 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259307 5763 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259461 5763 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 12:51:43.259545 5763 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"vices.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:51:45.973041 5964 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}\\\\nI1003 12:51:45.976192 5964 services_controller.go:444] Built service openshift-kube-apiserver-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1003 12:51:45.976195 5964 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1003 12:51:45.976169 5964 model_client.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.389966 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.402276 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.410222 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.410264 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.410278 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.410299 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.410310 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:54Z","lastTransitionTime":"2025-10-03T12:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.414904 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.425028 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.436355 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.447446 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.457350 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.470268 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.513216 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.513255 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.513266 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.513282 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.513293 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:54Z","lastTransitionTime":"2025-10-03T12:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.615199 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.615229 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.615237 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.615249 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.615257 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:54Z","lastTransitionTime":"2025-10-03T12:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.717321 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.717366 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.717377 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.717393 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.717404 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:54Z","lastTransitionTime":"2025-10-03T12:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.820108 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.820150 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.820160 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.820175 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.820185 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:54Z","lastTransitionTime":"2025-10-03T12:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.908527 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:54 crc kubenswrapper[4578]: E1003 12:51:54.908673 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.908784 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:51:54 crc kubenswrapper[4578]: E1003 12:51:54.908880 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.922444 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.922485 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.922494 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.922511 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.922521 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:54Z","lastTransitionTime":"2025-10-03T12:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.929039 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.940863 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.954711 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.974199 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b4650253599f9031287cc43642dff608baa3920e6dce60b433a7c642cece88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.988073 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:54 crc kubenswrapper[4578]: I1003 12:51:54.999813 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.019964 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"message\\\":\\\"work-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1003 12:51:43.258428 5763 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258471 5763 factory.go:656] Stopping watch factory\\\\nI1003 12:51:43.258513 5763 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258603 5763 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258755 5763 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259087 5763 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259307 5763 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259461 5763 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 12:51:43.259545 5763 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"vices.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:51:45.973041 5964 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}\\\\nI1003 12:51:45.976192 5964 services_controller.go:444] Built service openshift-kube-apiserver-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1003 12:51:45.976195 5964 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1003 12:51:45.976169 5964 model_client.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:55Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.023962 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.023989 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.023997 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.024008 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.024017 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:55Z","lastTransitionTime":"2025-10-03T12:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.029356 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:55Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.040005 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:55Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.050830 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:55Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.059500 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:55Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.070044 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:55Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.080787 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:55Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.089581 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:55Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.098504 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:55Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.125514 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.125544 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.125552 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.125566 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.125575 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:55Z","lastTransitionTime":"2025-10-03T12:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.228112 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.228470 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.228548 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.228614 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.228716 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:55Z","lastTransitionTime":"2025-10-03T12:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.331354 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.331393 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.331403 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.331418 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.331429 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:55Z","lastTransitionTime":"2025-10-03T12:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.434084 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.434451 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.434673 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.434830 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.434967 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:55Z","lastTransitionTime":"2025-10-03T12:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.538105 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.538810 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.538888 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.538976 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.539036 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:55Z","lastTransitionTime":"2025-10-03T12:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.640962 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.641228 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.641294 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.641358 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.641422 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:55Z","lastTransitionTime":"2025-10-03T12:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.744316 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.744591 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.744724 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.744809 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.744876 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:55Z","lastTransitionTime":"2025-10-03T12:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.846837 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.847136 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.847212 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.847289 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.847356 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:55Z","lastTransitionTime":"2025-10-03T12:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.908017 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.908152 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:55 crc kubenswrapper[4578]: E1003 12:51:55.908262 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:51:55 crc kubenswrapper[4578]: E1003 12:51:55.908427 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.949975 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.950012 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.950021 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.950036 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:55 crc kubenswrapper[4578]: I1003 12:51:55.950046 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:55Z","lastTransitionTime":"2025-10-03T12:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.052419 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.052479 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.052489 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.052570 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.052601 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:56Z","lastTransitionTime":"2025-10-03T12:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.158357 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.158397 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.158406 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.158420 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.158428 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:56Z","lastTransitionTime":"2025-10-03T12:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.260542 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.260582 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.260592 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.260607 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.260615 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:56Z","lastTransitionTime":"2025-10-03T12:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.362681 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.362984 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.363106 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.363199 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.363292 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:56Z","lastTransitionTime":"2025-10-03T12:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.465681 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.465718 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.465728 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.465744 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.465755 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:56Z","lastTransitionTime":"2025-10-03T12:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.568110 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.568139 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.568147 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.568159 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.568167 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:56Z","lastTransitionTime":"2025-10-03T12:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.670739 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.670776 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.670786 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.670802 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.670811 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:56Z","lastTransitionTime":"2025-10-03T12:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.773419 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.773761 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.773844 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.773936 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.774052 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:56Z","lastTransitionTime":"2025-10-03T12:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.876265 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.876309 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.876318 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.876335 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.876346 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:56Z","lastTransitionTime":"2025-10-03T12:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.908868 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:51:56 crc kubenswrapper[4578]: E1003 12:51:56.909052 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.908904 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:56 crc kubenswrapper[4578]: E1003 12:51:56.909672 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.978177 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.978465 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.978575 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.978691 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:56 crc kubenswrapper[4578]: I1003 12:51:56.978763 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:56Z","lastTransitionTime":"2025-10-03T12:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.065762 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.072457 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.077704 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:57Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.081276 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.081303 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.081310 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.081324 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.081333 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:57Z","lastTransitionTime":"2025-10-03T12:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.088983 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:57Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.101351 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:57Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.118398 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad9b3cc36c13b5ffc33760bb0994c16dedfecadf9cb83d0f143879f98286c3d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"message\\\":\\\"work-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1003 12:51:43.258428 5763 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258471 5763 factory.go:656] Stopping watch factory\\\\nI1003 12:51:43.258513 5763 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258603 5763 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.258755 5763 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259087 5763 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259307 5763 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1003 12:51:43.259461 5763 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1003 12:51:43.259545 5763 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"vices.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:51:45.973041 5964 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}\\\\nI1003 12:51:45.976192 5964 services_controller.go:444] Built service openshift-kube-apiserver-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1003 12:51:45.976195 5964 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1003 12:51:45.976169 5964 model_client.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:57Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.136726 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.136760 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.136781 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.136795 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.136806 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:57Z","lastTransitionTime":"2025-10-03T12:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:57 crc kubenswrapper[4578]: E1003 12:51:57.270230 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:57Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.270492 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:57Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.273704 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.273740 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.273756 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.273775 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.273785 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:57Z","lastTransitionTime":"2025-10-03T12:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.284456 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b4650253599f9031287cc43642dff608baa3920e6dce60b433a7c642cece88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:57Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:57 crc kubenswrapper[4578]: E1003 12:51:57.287207 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:57Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.291308 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.291340 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.291351 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.291368 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.291377 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:57Z","lastTransitionTime":"2025-10-03T12:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.296674 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:57Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:57 crc kubenswrapper[4578]: E1003 12:51:57.304009 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:57Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.306800 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:57Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.307001 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.307022 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.307030 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.307044 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.307052 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:57Z","lastTransitionTime":"2025-10-03T12:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.316080 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:57Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:57 crc kubenswrapper[4578]: E1003 12:51:57.318155 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:57Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.323348 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.323388 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.323396 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.323409 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.323419 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:57Z","lastTransitionTime":"2025-10-03T12:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.331213 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:57Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:57 crc kubenswrapper[4578]: E1003 12:51:57.339096 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:57Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:57 crc kubenswrapper[4578]: E1003 12:51:57.339279 4578 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.341135 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.341166 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.341174 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.341188 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.341198 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:57Z","lastTransitionTime":"2025-10-03T12:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.342901 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:57Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.355488 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:57Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.368725 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:57Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.382026 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:57Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.393447 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:57Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.444587 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.444666 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.444683 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.444701 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.444712 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:57Z","lastTransitionTime":"2025-10-03T12:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.548124 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.548182 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.548193 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.548209 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.548219 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:57Z","lastTransitionTime":"2025-10-03T12:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.651567 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.651703 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.651736 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.651771 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.651797 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:57Z","lastTransitionTime":"2025-10-03T12:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.754581 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.754645 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.754659 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.754676 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.754727 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:57Z","lastTransitionTime":"2025-10-03T12:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.857375 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.857414 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.857425 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.857439 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.857450 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:57Z","lastTransitionTime":"2025-10-03T12:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.908244 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:57 crc kubenswrapper[4578]: E1003 12:51:57.908378 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.908251 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:57 crc kubenswrapper[4578]: E1003 12:51:57.908521 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.959805 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.959885 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.959919 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.959935 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:57 crc kubenswrapper[4578]: I1003 12:51:57.959946 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:57Z","lastTransitionTime":"2025-10-03T12:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.062426 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.062466 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.062479 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.062494 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.062505 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:58Z","lastTransitionTime":"2025-10-03T12:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.164795 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.164856 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.164865 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.164880 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.164905 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:58Z","lastTransitionTime":"2025-10-03T12:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.267251 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.267293 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.267316 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.267334 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.267345 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:58Z","lastTransitionTime":"2025-10-03T12:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.369313 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.369364 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.369377 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.369393 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.369404 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:58Z","lastTransitionTime":"2025-10-03T12:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.471591 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.471625 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.471657 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.471671 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.471679 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:58Z","lastTransitionTime":"2025-10-03T12:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.574057 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.574381 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.574492 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.574595 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.574714 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:58Z","lastTransitionTime":"2025-10-03T12:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.677190 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.677784 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.677895 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.677993 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.678080 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:58Z","lastTransitionTime":"2025-10-03T12:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.780565 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.780604 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.780615 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.780654 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.780668 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:58Z","lastTransitionTime":"2025-10-03T12:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.883124 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.883378 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.883522 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.883612 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.883716 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:58Z","lastTransitionTime":"2025-10-03T12:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.908474 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:51:58 crc kubenswrapper[4578]: E1003 12:51:58.908599 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.908759 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:51:58 crc kubenswrapper[4578]: E1003 12:51:58.908807 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.986365 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.986749 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.986785 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.986806 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:58 crc kubenswrapper[4578]: I1003 12:51:58.986817 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:58Z","lastTransitionTime":"2025-10-03T12:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.090240 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.090279 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.090289 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.090303 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.090313 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:59Z","lastTransitionTime":"2025-10-03T12:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.192612 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.192672 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.192688 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.192738 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.192751 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:59Z","lastTransitionTime":"2025-10-03T12:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.295248 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.295287 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.295300 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.295317 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.295330 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:59Z","lastTransitionTime":"2025-10-03T12:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.400029 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.400063 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.400070 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.400083 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.400111 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:59Z","lastTransitionTime":"2025-10-03T12:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.502785 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.502832 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.502841 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.502855 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.502863 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:59Z","lastTransitionTime":"2025-10-03T12:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.605027 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.605081 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.605089 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.605105 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.605113 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:59Z","lastTransitionTime":"2025-10-03T12:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.691773 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.692681 4578 scope.go:117] "RemoveContainer" containerID="2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.702431 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:59Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.706940 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.707107 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.707176 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.707244 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.707303 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:59Z","lastTransitionTime":"2025-10-03T12:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.714491 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b4650253599f9031287cc43642dff608baa3920e6dce60b433a7c642cece88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:59Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.729722 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:59Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.744713 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:59Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.764978 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"vices.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:51:45.973041 5964 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}\\\\nI1003 12:51:45.976192 5964 services_controller.go:444] Built service openshift-kube-apiserver-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1003 12:51:45.976195 5964 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1003 12:51:45.976169 5964 model_client.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:59Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.779583 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:59Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.791615 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:59Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.803296 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:59Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.810485 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.810526 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.810537 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.810552 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.810563 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:59Z","lastTransitionTime":"2025-10-03T12:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.818710 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a524da53-3b46-423d-97ae-0e7d131b958f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d4f0ede4a9f9262ffe08a7abc4366faf1001b25f3a6a35dfac42c3d8362030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ba8c08b2dceaf712d8d6b27f5c3a4b6fd3a204fc494dfb8ec33ed7ce8bc16e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1fb8c53aa23e2c70af59f270c04d09599d79718098c17a809ace5f72797303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:59Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.833797 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:59Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.846656 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:59Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.856552 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:59Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.870772 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:59Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.883083 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:59Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.896108 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:59Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.908783 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.908783 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:51:59 crc kubenswrapper[4578]: E1003 12:51:59.908976 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:51:59 crc kubenswrapper[4578]: E1003 12:51:59.908905 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.911616 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:59Z is after 2025-08-24T17:21:41Z" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.912131 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.912156 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.912166 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.912180 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:51:59 crc kubenswrapper[4578]: I1003 12:51:59.912191 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:51:59Z","lastTransitionTime":"2025-10-03T12:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.014994 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.015046 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.015060 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.015079 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.015090 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:00Z","lastTransitionTime":"2025-10-03T12:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.123482 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.123518 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.123527 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.123542 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.123552 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:00Z","lastTransitionTime":"2025-10-03T12:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.226278 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.226317 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.226327 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.226342 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.226352 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:00Z","lastTransitionTime":"2025-10-03T12:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.294293 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8qvh9_ebaa96c5-be7c-4f05-819d-f37e1ceddac8/ovnkube-controller/1.log" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.296869 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerStarted","Data":"578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce"} Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.297798 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.314058 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.325239 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.329060 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.329110 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.329124 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.329147 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.329161 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:00Z","lastTransitionTime":"2025-10-03T12:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.340965 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.356321 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.370536 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.388802 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"vices.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:51:45.973041 5964 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}\\\\nI1003 12:51:45.976192 5964 services_controller.go:444] Built service openshift-kube-apiserver-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1003 12:51:45.976195 5964 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1003 12:51:45.976169 5964 model_client.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.400245 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.417084 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b4650253599f9031287cc43642dff608baa3920e6dce60b433a7c642cece88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.432061 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.432104 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.432116 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.432130 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.432141 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:00Z","lastTransitionTime":"2025-10-03T12:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.432859 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.446497 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.462487 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.475253 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.490008 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.504167 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.517251 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a524da53-3b46-423d-97ae-0e7d131b958f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d4f0ede4a9f9262ffe08a7abc4366faf1001b25f3a6a35dfac42c3d8362030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ba8c08b2dceaf712d8d6b27f5c3a4b6fd3a204fc494dfb8ec33ed7ce8bc16e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1fb8c53aa23e2c70af59f270c04d09599d79718098c17a809ace5f72797303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.529260 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.535269 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.535314 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.535324 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.535342 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.535354 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:00Z","lastTransitionTime":"2025-10-03T12:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.637522 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.637605 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.637621 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.637682 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.637692 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:00Z","lastTransitionTime":"2025-10-03T12:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.740578 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.740613 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.740622 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.740657 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.740666 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:00Z","lastTransitionTime":"2025-10-03T12:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.843251 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.843288 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.843296 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.843316 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.843327 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:00Z","lastTransitionTime":"2025-10-03T12:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.909077 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:00 crc kubenswrapper[4578]: E1003 12:52:00.909198 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.909081 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:00 crc kubenswrapper[4578]: E1003 12:52:00.909377 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.945938 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.945980 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.945989 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.946002 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:00 crc kubenswrapper[4578]: I1003 12:52:00.946011 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:00Z","lastTransitionTime":"2025-10-03T12:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.048245 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.048277 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.048285 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.048297 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.048306 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:01Z","lastTransitionTime":"2025-10-03T12:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.149886 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.149917 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.149927 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.149943 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.149953 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:01Z","lastTransitionTime":"2025-10-03T12:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.188188 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs\") pod \"network-metrics-daemon-q8986\" (UID: \"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\") " pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:01 crc kubenswrapper[4578]: E1003 12:52:01.188286 4578 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 12:52:01 crc kubenswrapper[4578]: E1003 12:52:01.188328 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs podName:2971c7bb-e0f4-44a8-8f52-9ea8ba199772 nodeName:}" failed. No retries permitted until 2025-10-03 12:52:17.188315558 +0000 UTC m=+72.986787732 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs") pod "network-metrics-daemon-q8986" (UID: "2971c7bb-e0f4-44a8-8f52-9ea8ba199772") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.253036 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.253087 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.253098 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.253114 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.253123 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:01Z","lastTransitionTime":"2025-10-03T12:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.303787 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8qvh9_ebaa96c5-be7c-4f05-819d-f37e1ceddac8/ovnkube-controller/2.log" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.304604 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8qvh9_ebaa96c5-be7c-4f05-819d-f37e1ceddac8/ovnkube-controller/1.log" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.308309 4578 generic.go:334] "Generic (PLEG): container finished" podID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerID="578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce" exitCode=1 Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.308359 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerDied","Data":"578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce"} Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.308396 4578 scope.go:117] "RemoveContainer" containerID="2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.309032 4578 scope.go:117] "RemoveContainer" containerID="578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce" Oct 03 12:52:01 crc kubenswrapper[4578]: E1003 12:52:01.309203 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.323903 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:01Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.336048 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:01Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.346820 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:01Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.356303 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.356334 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.356342 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.356355 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.356366 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:01Z","lastTransitionTime":"2025-10-03T12:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.360155 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a524da53-3b46-423d-97ae-0e7d131b958f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d4f0ede4a9f9262ffe08a7abc4366faf1001b25f3a6a35dfac42c3d8362030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ba8c08b2dceaf712d8d6b27f5c3a4b6fd3a204fc494dfb8ec33ed7ce8bc16e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1fb8c53aa23e2c70af59f270c04d09599d79718098c17a809ace5f72797303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:01Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.376607 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:01Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.389698 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:01Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.401456 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:01Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.412169 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:01Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.426869 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:01Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.440330 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:01Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.458319 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:01Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.458885 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.458960 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.458978 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.459005 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.459021 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:01Z","lastTransitionTime":"2025-10-03T12:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.473867 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b4650253599f9031287cc43642dff608baa3920e6dce60b433a7c642cece88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:01Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.491846 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:01Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.503872 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:01Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.521311 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f9d2ea003a5f42376da2914e5bd5ae580d66e1ce02ee07e1af8246f2a102ed5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"vices.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:51:45.973041 5964 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}\\\\nI1003 12:51:45.976192 5964 services_controller.go:444] Built service openshift-kube-apiserver-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1003 12:51:45.976195 5964 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1003 12:51:45.976169 5964 model_client.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:00Z\\\",\\\"message\\\":\\\"rvices.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:52:00.501792 6175 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nF1003 12:52:00.501828 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z]\\\\nI1003 12:52:00.501838 6175 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nI1003 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:01Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.532116 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:01Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.563439 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.563513 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.563530 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.563559 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.563584 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:01Z","lastTransitionTime":"2025-10-03T12:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.666617 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.666686 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.666697 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.666713 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.666723 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:01Z","lastTransitionTime":"2025-10-03T12:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.768885 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.768915 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.768922 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.768936 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.768946 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:01Z","lastTransitionTime":"2025-10-03T12:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.871482 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.871521 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.871529 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.871542 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.871551 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:01Z","lastTransitionTime":"2025-10-03T12:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.908342 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.908342 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:01 crc kubenswrapper[4578]: E1003 12:52:01.908489 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:01 crc kubenswrapper[4578]: E1003 12:52:01.908544 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.973699 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.973734 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.973743 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.973757 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:01 crc kubenswrapper[4578]: I1003 12:52:01.973767 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:01Z","lastTransitionTime":"2025-10-03T12:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.076054 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.076319 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.076390 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.076478 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.076542 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:02Z","lastTransitionTime":"2025-10-03T12:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.179655 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.179698 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.179707 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.179724 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.179734 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:02Z","lastTransitionTime":"2025-10-03T12:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.282350 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.282393 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.282404 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.282417 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.282426 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:02Z","lastTransitionTime":"2025-10-03T12:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.314304 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8qvh9_ebaa96c5-be7c-4f05-819d-f37e1ceddac8/ovnkube-controller/2.log" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.319053 4578 scope.go:117] "RemoveContainer" containerID="578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce" Oct 03 12:52:02 crc kubenswrapper[4578]: E1003 12:52:02.319537 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.337068 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:02Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.358773 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:00Z\\\",\\\"message\\\":\\\"rvices.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:52:00.501792 6175 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nF1003 12:52:00.501828 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z]\\\\nI1003 12:52:00.501838 6175 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nI1003 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:02Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.370821 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:02Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.384659 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.384694 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.384701 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.384716 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.384724 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:02Z","lastTransitionTime":"2025-10-03T12:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.394755 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b4650253599f9031287cc43642dff608baa3920e6dce60b433a7c642cece88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:02Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.408536 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:02Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.419924 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:02Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.429372 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:02Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.442175 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:02Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.451966 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:02Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.461704 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:02Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.475319 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a524da53-3b46-423d-97ae-0e7d131b958f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d4f0ede4a9f9262ffe08a7abc4366faf1001b25f3a6a35dfac42c3d8362030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ba8c08b2dceaf712d8d6b27f5c3a4b6fd3a204fc494dfb8ec33ed7ce8bc16e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1fb8c53aa23e2c70af59f270c04d09599d79718098c17a809ace5f72797303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:02Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.486726 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:02Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.487132 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.487174 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.487183 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.487199 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.487210 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:02Z","lastTransitionTime":"2025-10-03T12:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.502112 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:02Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.519677 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:02Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.530811 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:02Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.544310 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:02Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.590109 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.590145 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.590155 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.590199 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.590210 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:02Z","lastTransitionTime":"2025-10-03T12:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.692341 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.692382 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.692391 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.692405 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.692419 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:02Z","lastTransitionTime":"2025-10-03T12:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.795212 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.795261 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.795274 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.795293 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.795308 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:02Z","lastTransitionTime":"2025-10-03T12:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.897364 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.897420 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.897433 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.897451 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.897462 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:02Z","lastTransitionTime":"2025-10-03T12:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.909109 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:02 crc kubenswrapper[4578]: I1003 12:52:02.909164 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:02 crc kubenswrapper[4578]: E1003 12:52:02.909243 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:02 crc kubenswrapper[4578]: E1003 12:52:02.909345 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.000242 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.000324 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.000353 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.000385 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.000405 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:03Z","lastTransitionTime":"2025-10-03T12:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.102995 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.103063 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.103086 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.103124 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.103143 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:03Z","lastTransitionTime":"2025-10-03T12:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.205393 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.205455 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.205472 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.205490 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.205527 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:03Z","lastTransitionTime":"2025-10-03T12:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.308708 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.308778 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.308790 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.308807 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.308820 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:03Z","lastTransitionTime":"2025-10-03T12:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.411512 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.411609 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.411618 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.411666 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.411683 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:03Z","lastTransitionTime":"2025-10-03T12:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.514156 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.514205 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.514218 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.514237 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.514250 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:03Z","lastTransitionTime":"2025-10-03T12:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.616562 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.616603 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.616612 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.616663 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.616674 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:03Z","lastTransitionTime":"2025-10-03T12:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.707943 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:52:03 crc kubenswrapper[4578]: E1003 12:52:03.708130 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:52:35.708096623 +0000 UTC m=+91.506568817 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.719651 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.719701 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.719714 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.719731 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.719743 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:03Z","lastTransitionTime":"2025-10-03T12:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.809346 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.809456 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.809518 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.809583 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:03 crc kubenswrapper[4578]: E1003 12:52:03.809616 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 12:52:03 crc kubenswrapper[4578]: E1003 12:52:03.809670 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 12:52:03 crc kubenswrapper[4578]: E1003 12:52:03.809684 4578 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:52:03 crc kubenswrapper[4578]: E1003 12:52:03.809740 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 12:52:35.809720231 +0000 UTC m=+91.608192415 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:52:03 crc kubenswrapper[4578]: E1003 12:52:03.809747 4578 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 12:52:03 crc kubenswrapper[4578]: E1003 12:52:03.809761 4578 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 12:52:03 crc kubenswrapper[4578]: E1003 12:52:03.809790 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 12:52:35.809779104 +0000 UTC m=+91.608251288 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 12:52:03 crc kubenswrapper[4578]: E1003 12:52:03.809867 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 12:52:35.809837425 +0000 UTC m=+91.608309629 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 12:52:03 crc kubenswrapper[4578]: E1003 12:52:03.810152 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 12:52:03 crc kubenswrapper[4578]: E1003 12:52:03.810208 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 12:52:03 crc kubenswrapper[4578]: E1003 12:52:03.810225 4578 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:52:03 crc kubenswrapper[4578]: E1003 12:52:03.810299 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 12:52:35.810279719 +0000 UTC m=+91.608751903 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.822492 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.822549 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.822557 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.822599 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.822609 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:03Z","lastTransitionTime":"2025-10-03T12:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.908577 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.908686 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:03 crc kubenswrapper[4578]: E1003 12:52:03.908746 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:03 crc kubenswrapper[4578]: E1003 12:52:03.908889 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.924573 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.924665 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.924684 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.924703 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:03 crc kubenswrapper[4578]: I1003 12:52:03.924717 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:03Z","lastTransitionTime":"2025-10-03T12:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.026868 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.026899 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.026908 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.026920 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.026929 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:04Z","lastTransitionTime":"2025-10-03T12:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.130097 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.130144 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.130155 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.130172 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.130182 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:04Z","lastTransitionTime":"2025-10-03T12:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.232705 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.232747 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.232757 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.232773 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.232785 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:04Z","lastTransitionTime":"2025-10-03T12:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.334699 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.334788 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.334803 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.334830 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.334844 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:04Z","lastTransitionTime":"2025-10-03T12:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.437132 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.437176 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.437187 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.437200 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.437209 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:04Z","lastTransitionTime":"2025-10-03T12:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.539282 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.539341 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.539352 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.539366 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.539377 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:04Z","lastTransitionTime":"2025-10-03T12:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.642011 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.642060 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.642071 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.642087 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.642101 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:04Z","lastTransitionTime":"2025-10-03T12:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.744015 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.744054 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.744064 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.744079 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.744087 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:04Z","lastTransitionTime":"2025-10-03T12:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.846296 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.846331 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.846345 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.846360 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.846370 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:04Z","lastTransitionTime":"2025-10-03T12:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.908375 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:04 crc kubenswrapper[4578]: E1003 12:52:04.908500 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.910586 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:04 crc kubenswrapper[4578]: E1003 12:52:04.910753 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.925076 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:04Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.937535 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:04Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.948186 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:04Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.949789 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.949825 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.949836 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.949853 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.949863 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:04Z","lastTransitionTime":"2025-10-03T12:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.960265 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:04Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.973192 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a524da53-3b46-423d-97ae-0e7d131b958f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d4f0ede4a9f9262ffe08a7abc4366faf1001b25f3a6a35dfac42c3d8362030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ba8c08b2dceaf712d8d6b27f5c3a4b6fd3a204fc494dfb8ec33ed7ce8bc16e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1fb8c53aa23e2c70af59f270c04d09599d79718098c17a809ace5f72797303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:04Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:04 crc kubenswrapper[4578]: I1003 12:52:04.986728 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:04Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.003977 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:05Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.019506 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:05Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.032801 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:05Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.044761 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:05Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.052594 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.052639 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.052650 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.052665 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.052676 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:05Z","lastTransitionTime":"2025-10-03T12:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.065329 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:00Z\\\",\\\"message\\\":\\\"rvices.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:52:00.501792 6175 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nF1003 12:52:00.501828 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z]\\\\nI1003 12:52:00.501838 6175 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nI1003 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:05Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.076061 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:05Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.088686 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b4650253599f9031287cc43642dff608baa3920e6dce60b433a7c642cece88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:05Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.100721 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:05Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.111579 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:05Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.122262 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:05Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.154449 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.154490 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.154509 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.154525 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.154679 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:05Z","lastTransitionTime":"2025-10-03T12:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.257411 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.257455 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.257470 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.257486 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.257497 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:05Z","lastTransitionTime":"2025-10-03T12:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.360147 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.360759 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.360842 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.360933 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.361018 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:05Z","lastTransitionTime":"2025-10-03T12:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.463686 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.463725 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.463734 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.463751 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.463760 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:05Z","lastTransitionTime":"2025-10-03T12:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.565857 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.565891 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.565900 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.565914 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.565923 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:05Z","lastTransitionTime":"2025-10-03T12:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.668061 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.668124 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.668134 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.668148 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.668158 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:05Z","lastTransitionTime":"2025-10-03T12:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.770498 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.770537 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.770549 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.770564 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.770574 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:05Z","lastTransitionTime":"2025-10-03T12:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.872868 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.872904 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.872913 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.872927 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.872962 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:05Z","lastTransitionTime":"2025-10-03T12:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.908438 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.908595 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:05 crc kubenswrapper[4578]: E1003 12:52:05.908726 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:05 crc kubenswrapper[4578]: E1003 12:52:05.908879 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.975264 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.975325 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.975336 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.975352 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:05 crc kubenswrapper[4578]: I1003 12:52:05.975362 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:05Z","lastTransitionTime":"2025-10-03T12:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.078086 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.078126 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.078134 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.078147 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.078158 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:06Z","lastTransitionTime":"2025-10-03T12:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.181221 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.181280 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.181291 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.181305 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.181318 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:06Z","lastTransitionTime":"2025-10-03T12:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.284534 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.284602 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.284626 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.284717 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.284743 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:06Z","lastTransitionTime":"2025-10-03T12:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.386852 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.386886 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.386895 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.386911 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.386923 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:06Z","lastTransitionTime":"2025-10-03T12:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.491462 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.491555 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.491571 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.491610 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.491624 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:06Z","lastTransitionTime":"2025-10-03T12:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.594051 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.594090 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.594122 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.594141 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.594153 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:06Z","lastTransitionTime":"2025-10-03T12:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.696183 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.696220 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.696232 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.696246 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.696257 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:06Z","lastTransitionTime":"2025-10-03T12:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.798118 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.798163 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.798174 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.798190 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.798200 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:06Z","lastTransitionTime":"2025-10-03T12:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.901730 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.902617 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.902845 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.902984 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.903254 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:06Z","lastTransitionTime":"2025-10-03T12:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.909007 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:06 crc kubenswrapper[4578]: I1003 12:52:06.909104 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:06 crc kubenswrapper[4578]: E1003 12:52:06.909214 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:06 crc kubenswrapper[4578]: E1003 12:52:06.909324 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.006291 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.006549 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.006773 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.006934 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.007060 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:07Z","lastTransitionTime":"2025-10-03T12:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.110669 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.110730 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.110743 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.110758 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.110768 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:07Z","lastTransitionTime":"2025-10-03T12:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.213270 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.213310 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.213322 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.213337 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.213348 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:07Z","lastTransitionTime":"2025-10-03T12:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.316325 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.316389 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.316401 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.316417 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.316433 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:07Z","lastTransitionTime":"2025-10-03T12:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.418290 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.418330 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.418338 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.418356 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.418367 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:07Z","lastTransitionTime":"2025-10-03T12:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.506970 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.507019 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.507034 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.507051 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.507062 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:07Z","lastTransitionTime":"2025-10-03T12:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:07 crc kubenswrapper[4578]: E1003 12:52:07.520301 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:07Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.523670 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.523711 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.523752 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.523768 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.523779 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:07Z","lastTransitionTime":"2025-10-03T12:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:07 crc kubenswrapper[4578]: E1003 12:52:07.538836 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:07Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.542989 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.543023 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.543033 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.543049 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.543057 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:07Z","lastTransitionTime":"2025-10-03T12:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:07 crc kubenswrapper[4578]: E1003 12:52:07.557744 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:07Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.561173 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.561203 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.561213 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.561226 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.561235 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:07Z","lastTransitionTime":"2025-10-03T12:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:07 crc kubenswrapper[4578]: E1003 12:52:07.572912 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:07Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.576536 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.576584 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.576595 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.576611 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.576620 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:07Z","lastTransitionTime":"2025-10-03T12:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:07 crc kubenswrapper[4578]: E1003 12:52:07.587338 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:07Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:07 crc kubenswrapper[4578]: E1003 12:52:07.587530 4578 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.589162 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.589191 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.589200 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.589449 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.589472 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:07Z","lastTransitionTime":"2025-10-03T12:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.692967 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.693006 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.693018 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.693056 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.693068 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:07Z","lastTransitionTime":"2025-10-03T12:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.795831 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.795902 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.795916 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.795934 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.795968 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:07Z","lastTransitionTime":"2025-10-03T12:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.899183 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.899226 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.899236 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.899250 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.899260 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:07Z","lastTransitionTime":"2025-10-03T12:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.908487 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:07 crc kubenswrapper[4578]: I1003 12:52:07.908487 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:07 crc kubenswrapper[4578]: E1003 12:52:07.908657 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:07 crc kubenswrapper[4578]: E1003 12:52:07.908702 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.002149 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.002203 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.002216 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.002232 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.002262 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:08Z","lastTransitionTime":"2025-10-03T12:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.105074 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.105121 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.105130 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.105147 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.105158 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:08Z","lastTransitionTime":"2025-10-03T12:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.207810 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.207859 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.207870 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.207884 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.207893 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:08Z","lastTransitionTime":"2025-10-03T12:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.310196 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.310236 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.310246 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.310260 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.310269 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:08Z","lastTransitionTime":"2025-10-03T12:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.413864 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.413903 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.413912 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.413926 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.413935 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:08Z","lastTransitionTime":"2025-10-03T12:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.516392 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.516429 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.516437 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.516450 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.516461 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:08Z","lastTransitionTime":"2025-10-03T12:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.618564 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.618606 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.618617 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.618651 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.618666 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:08Z","lastTransitionTime":"2025-10-03T12:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.721284 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.721326 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.721337 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.721353 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.721364 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:08Z","lastTransitionTime":"2025-10-03T12:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.824234 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.824285 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.824297 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.824315 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.824328 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:08Z","lastTransitionTime":"2025-10-03T12:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.908573 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.908662 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:08 crc kubenswrapper[4578]: E1003 12:52:08.908944 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:08 crc kubenswrapper[4578]: E1003 12:52:08.909024 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.927200 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.927246 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.927255 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.927271 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:08 crc kubenswrapper[4578]: I1003 12:52:08.927280 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:08Z","lastTransitionTime":"2025-10-03T12:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.029275 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.029353 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.029363 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.029377 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.029387 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:09Z","lastTransitionTime":"2025-10-03T12:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.101896 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.116349 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:09Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.127883 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:09Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.131444 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.131495 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.131504 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.131517 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.131528 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:09Z","lastTransitionTime":"2025-10-03T12:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.138506 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:09Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.150163 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a524da53-3b46-423d-97ae-0e7d131b958f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d4f0ede4a9f9262ffe08a7abc4366faf1001b25f3a6a35dfac42c3d8362030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ba8c08b2dceaf712d8d6b27f5c3a4b6fd3a204fc494dfb8ec33ed7ce8bc16e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1fb8c53aa23e2c70af59f270c04d09599d79718098c17a809ace5f72797303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:09Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.165087 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:09Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.176754 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:09Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.188600 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:09Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.199126 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:09Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.209825 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:09Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.221342 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:09Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.233705 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.233748 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.233756 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.233770 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.233780 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:09Z","lastTransitionTime":"2025-10-03T12:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.234855 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:09Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.248284 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b4650253599f9031287cc43642dff608baa3920e6dce60b433a7c642cece88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:09Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.260556 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:09Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.274563 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:09Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.292511 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:00Z\\\",\\\"message\\\":\\\"rvices.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:52:00.501792 6175 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nF1003 12:52:00.501828 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z]\\\\nI1003 12:52:00.501838 6175 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nI1003 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:09Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.302759 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:09Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.335849 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.335889 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.335902 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.335918 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.335931 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:09Z","lastTransitionTime":"2025-10-03T12:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.438514 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.438545 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.438553 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.438566 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.438574 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:09Z","lastTransitionTime":"2025-10-03T12:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.540819 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.540883 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.540894 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.540914 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.540925 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:09Z","lastTransitionTime":"2025-10-03T12:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.643804 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.643849 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.643858 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.643875 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.643885 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:09Z","lastTransitionTime":"2025-10-03T12:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.746133 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.746164 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.746172 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.746185 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.746193 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:09Z","lastTransitionTime":"2025-10-03T12:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.849509 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.849562 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.849573 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.849586 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.849610 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:09Z","lastTransitionTime":"2025-10-03T12:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.908245 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.908286 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:09 crc kubenswrapper[4578]: E1003 12:52:09.908388 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:09 crc kubenswrapper[4578]: E1003 12:52:09.908469 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.952350 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.952388 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.952396 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.952411 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:09 crc kubenswrapper[4578]: I1003 12:52:09.952420 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:09Z","lastTransitionTime":"2025-10-03T12:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.055354 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.055393 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.055404 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.055420 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.055431 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:10Z","lastTransitionTime":"2025-10-03T12:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.158132 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.158178 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.158208 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.158223 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.158234 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:10Z","lastTransitionTime":"2025-10-03T12:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.260787 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.260834 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.260846 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.260864 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.260875 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:10Z","lastTransitionTime":"2025-10-03T12:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.363913 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.363977 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.364015 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.364039 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.364056 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:10Z","lastTransitionTime":"2025-10-03T12:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.466620 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.466706 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.466721 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.466746 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.466762 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:10Z","lastTransitionTime":"2025-10-03T12:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.569670 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.569708 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.569716 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.569730 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.569739 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:10Z","lastTransitionTime":"2025-10-03T12:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.672525 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.672569 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.672580 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.672599 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.672756 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:10Z","lastTransitionTime":"2025-10-03T12:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.775138 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.775209 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.775220 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.775235 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.775246 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:10Z","lastTransitionTime":"2025-10-03T12:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.878404 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.878474 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.878488 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.878502 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.878513 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:10Z","lastTransitionTime":"2025-10-03T12:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.908769 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.908873 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:10 crc kubenswrapper[4578]: E1003 12:52:10.908897 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:10 crc kubenswrapper[4578]: E1003 12:52:10.909055 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.981271 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.981308 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.981318 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.981332 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:10 crc kubenswrapper[4578]: I1003 12:52:10.981342 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:10Z","lastTransitionTime":"2025-10-03T12:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.087320 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.087364 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.087375 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.087390 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.087399 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:11Z","lastTransitionTime":"2025-10-03T12:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.190250 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.190293 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.190305 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.190322 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.190333 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:11Z","lastTransitionTime":"2025-10-03T12:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.295344 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.295376 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.295386 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.295400 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.295409 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:11Z","lastTransitionTime":"2025-10-03T12:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.398196 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.398240 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.398251 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.398268 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.398281 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:11Z","lastTransitionTime":"2025-10-03T12:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.501300 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.501353 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.501363 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.501379 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.501391 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:11Z","lastTransitionTime":"2025-10-03T12:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.603879 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.604068 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.604083 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.604098 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.604108 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:11Z","lastTransitionTime":"2025-10-03T12:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.706140 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.706175 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.706183 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.706199 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.706207 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:11Z","lastTransitionTime":"2025-10-03T12:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.808327 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.808366 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.808374 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.808388 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.808397 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:11Z","lastTransitionTime":"2025-10-03T12:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.908863 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.908922 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:11 crc kubenswrapper[4578]: E1003 12:52:11.909082 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:11 crc kubenswrapper[4578]: E1003 12:52:11.909008 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.910424 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.910472 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.910483 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.910494 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:11 crc kubenswrapper[4578]: I1003 12:52:11.910504 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:11Z","lastTransitionTime":"2025-10-03T12:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.012763 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.012812 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.012822 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.012838 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.012851 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:12Z","lastTransitionTime":"2025-10-03T12:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.115342 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.115377 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.115385 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.115400 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.115408 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:12Z","lastTransitionTime":"2025-10-03T12:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.217430 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.217483 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.217492 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.217504 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.217513 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:12Z","lastTransitionTime":"2025-10-03T12:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.319919 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.319967 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.320009 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.320026 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.320035 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:12Z","lastTransitionTime":"2025-10-03T12:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.423374 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.423428 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.423437 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.423455 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.423466 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:12Z","lastTransitionTime":"2025-10-03T12:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.526129 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.526168 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.526179 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.526194 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.526205 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:12Z","lastTransitionTime":"2025-10-03T12:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.629000 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.629052 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.629064 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.629080 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.629089 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:12Z","lastTransitionTime":"2025-10-03T12:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.731665 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.731703 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.731712 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.731728 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.731737 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:12Z","lastTransitionTime":"2025-10-03T12:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.834047 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.834101 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.834114 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.834128 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.834137 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:12Z","lastTransitionTime":"2025-10-03T12:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.908852 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.908986 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:12 crc kubenswrapper[4578]: E1003 12:52:12.909139 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:12 crc kubenswrapper[4578]: E1003 12:52:12.909330 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.937128 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.937172 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.937183 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.937198 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:12 crc kubenswrapper[4578]: I1003 12:52:12.937210 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:12Z","lastTransitionTime":"2025-10-03T12:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.039614 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.039662 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.039669 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.039684 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.039692 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:13Z","lastTransitionTime":"2025-10-03T12:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.143163 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.143220 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.143231 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.143265 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.143282 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:13Z","lastTransitionTime":"2025-10-03T12:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.245802 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.245856 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.245867 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.245883 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.245912 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:13Z","lastTransitionTime":"2025-10-03T12:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.347462 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.347514 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.347528 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.347549 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.347561 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:13Z","lastTransitionTime":"2025-10-03T12:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.450340 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.450399 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.450414 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.450459 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.450470 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:13Z","lastTransitionTime":"2025-10-03T12:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.553611 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.553689 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.553699 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.553715 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.553725 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:13Z","lastTransitionTime":"2025-10-03T12:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.656007 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.656072 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.656081 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.656105 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.656115 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:13Z","lastTransitionTime":"2025-10-03T12:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.758505 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.758573 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.758586 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.758599 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.758608 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:13Z","lastTransitionTime":"2025-10-03T12:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.861239 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.861283 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.861294 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.861309 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.861320 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:13Z","lastTransitionTime":"2025-10-03T12:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.908847 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:13 crc kubenswrapper[4578]: E1003 12:52:13.908997 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.909162 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:13 crc kubenswrapper[4578]: E1003 12:52:13.909205 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.963518 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.963567 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.963577 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.963594 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:13 crc kubenswrapper[4578]: I1003 12:52:13.963605 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:13Z","lastTransitionTime":"2025-10-03T12:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.065667 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.065706 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.065717 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.065732 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.065753 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:14Z","lastTransitionTime":"2025-10-03T12:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.167907 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.167969 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.167980 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.168009 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.168021 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:14Z","lastTransitionTime":"2025-10-03T12:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.271071 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.271098 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.271106 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.271119 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.271127 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:14Z","lastTransitionTime":"2025-10-03T12:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.373964 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.374001 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.374013 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.374029 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.374040 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:14Z","lastTransitionTime":"2025-10-03T12:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.476291 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.476325 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.476336 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.476352 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.476363 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:14Z","lastTransitionTime":"2025-10-03T12:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.578774 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.578815 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.578827 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.578844 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.578857 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:14Z","lastTransitionTime":"2025-10-03T12:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.681121 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.681158 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.681170 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.681186 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.681196 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:14Z","lastTransitionTime":"2025-10-03T12:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.784384 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.784441 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.784453 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.784470 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.784482 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:14Z","lastTransitionTime":"2025-10-03T12:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.887038 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.887087 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.887098 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.887117 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.887131 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:14Z","lastTransitionTime":"2025-10-03T12:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.908477 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:14 crc kubenswrapper[4578]: E1003 12:52:14.908611 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.908652 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:14 crc kubenswrapper[4578]: E1003 12:52:14.908775 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.922855 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:14Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.933496 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:14Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.945780 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:14Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.958304 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:14Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.969709 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:14Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.981660 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:14Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.989416 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.989458 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.989467 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.989480 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.989490 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:14Z","lastTransitionTime":"2025-10-03T12:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:14 crc kubenswrapper[4578]: I1003 12:52:14.994075 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a524da53-3b46-423d-97ae-0e7d131b958f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d4f0ede4a9f9262ffe08a7abc4366faf1001b25f3a6a35dfac42c3d8362030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ba8c08b2dceaf712d8d6b27f5c3a4b6fd3a204fc494dfb8ec33ed7ce8bc16e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1fb8c53aa23e2c70af59f270c04d09599d79718098c17a809ace5f72797303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:14Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.008752 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:15Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.023011 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:15Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.033621 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:15Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.043622 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:15Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.055305 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:15Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.066700 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:15Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.084030 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:00Z\\\",\\\"message\\\":\\\"rvices.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:52:00.501792 6175 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nF1003 12:52:00.501828 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z]\\\\nI1003 12:52:00.501838 6175 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nI1003 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:15Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.091511 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.091549 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.091561 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.091575 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.091586 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:15Z","lastTransitionTime":"2025-10-03T12:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.096291 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:15Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.108846 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b4650253599f9031287cc43642dff608baa3920e6dce60b433a7c642cece88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:15Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.193881 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.193923 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.193932 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.193948 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.193957 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:15Z","lastTransitionTime":"2025-10-03T12:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.296760 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.296804 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.296813 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.296828 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.296837 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:15Z","lastTransitionTime":"2025-10-03T12:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.399382 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.399429 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.399440 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.399458 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.399470 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:15Z","lastTransitionTime":"2025-10-03T12:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.501207 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.501236 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.501260 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.501275 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.501284 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:15Z","lastTransitionTime":"2025-10-03T12:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.603901 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.603960 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.603978 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.604002 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.604020 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:15Z","lastTransitionTime":"2025-10-03T12:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.706604 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.706664 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.706677 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.706779 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.706793 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:15Z","lastTransitionTime":"2025-10-03T12:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.808504 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.808536 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.808544 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.808556 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.808564 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:15Z","lastTransitionTime":"2025-10-03T12:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.908540 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:15 crc kubenswrapper[4578]: E1003 12:52:15.908705 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.908919 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:15 crc kubenswrapper[4578]: E1003 12:52:15.908989 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.910671 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.910699 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.910708 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.910731 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:15 crc kubenswrapper[4578]: I1003 12:52:15.910740 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:15Z","lastTransitionTime":"2025-10-03T12:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.012814 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.012851 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.012859 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.012873 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.012882 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:16Z","lastTransitionTime":"2025-10-03T12:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.115094 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.115139 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.115151 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.115166 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.115178 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:16Z","lastTransitionTime":"2025-10-03T12:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.217515 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.217564 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.217575 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.217592 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.217603 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:16Z","lastTransitionTime":"2025-10-03T12:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.319936 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.319979 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.320036 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.320056 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.320069 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:16Z","lastTransitionTime":"2025-10-03T12:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.422608 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.422671 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.422682 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.422699 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.422710 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:16Z","lastTransitionTime":"2025-10-03T12:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.524801 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.524841 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.524852 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.524867 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.524877 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:16Z","lastTransitionTime":"2025-10-03T12:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.627152 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.627191 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.627201 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.627217 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.627228 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:16Z","lastTransitionTime":"2025-10-03T12:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.729175 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.729242 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.729254 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.729288 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.729301 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:16Z","lastTransitionTime":"2025-10-03T12:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.831607 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.831676 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.831688 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.831706 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.831718 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:16Z","lastTransitionTime":"2025-10-03T12:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.909109 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.909146 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:16 crc kubenswrapper[4578]: E1003 12:52:16.909528 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:16 crc kubenswrapper[4578]: E1003 12:52:16.909654 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.909816 4578 scope.go:117] "RemoveContainer" containerID="578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce" Oct 03 12:52:16 crc kubenswrapper[4578]: E1003 12:52:16.909967 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.933513 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.933582 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.933593 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.933606 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:16 crc kubenswrapper[4578]: I1003 12:52:16.933617 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:16Z","lastTransitionTime":"2025-10-03T12:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.036311 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.036345 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.036356 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.036372 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.036381 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:17Z","lastTransitionTime":"2025-10-03T12:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.138794 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.138835 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.138846 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.138861 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.138873 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:17Z","lastTransitionTime":"2025-10-03T12:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.240865 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.240909 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.240920 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.240935 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.240944 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:17Z","lastTransitionTime":"2025-10-03T12:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.245179 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs\") pod \"network-metrics-daemon-q8986\" (UID: \"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\") " pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:17 crc kubenswrapper[4578]: E1003 12:52:17.245310 4578 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 12:52:17 crc kubenswrapper[4578]: E1003 12:52:17.245361 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs podName:2971c7bb-e0f4-44a8-8f52-9ea8ba199772 nodeName:}" failed. No retries permitted until 2025-10-03 12:52:49.245348446 +0000 UTC m=+105.043820630 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs") pod "network-metrics-daemon-q8986" (UID: "2971c7bb-e0f4-44a8-8f52-9ea8ba199772") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.342778 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.342820 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.342831 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.342849 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.342861 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:17Z","lastTransitionTime":"2025-10-03T12:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.445662 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.445707 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.445718 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.445735 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.445746 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:17Z","lastTransitionTime":"2025-10-03T12:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.548280 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.548337 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.548349 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.548367 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.548383 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:17Z","lastTransitionTime":"2025-10-03T12:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.650668 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.650717 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.650729 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.650745 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.650757 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:17Z","lastTransitionTime":"2025-10-03T12:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.753670 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.753718 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.753729 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.753745 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.753756 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:17Z","lastTransitionTime":"2025-10-03T12:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.856524 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.856562 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.856570 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.856582 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.856591 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:17Z","lastTransitionTime":"2025-10-03T12:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.899439 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.899499 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.899515 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.899537 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.899553 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:17Z","lastTransitionTime":"2025-10-03T12:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.908449 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:17 crc kubenswrapper[4578]: E1003 12:52:17.908593 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.908806 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:17 crc kubenswrapper[4578]: E1003 12:52:17.908855 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:17 crc kubenswrapper[4578]: E1003 12:52:17.918178 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:17Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.921436 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.921460 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.921484 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.921497 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.921506 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:17Z","lastTransitionTime":"2025-10-03T12:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:17 crc kubenswrapper[4578]: E1003 12:52:17.934341 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:17Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.937244 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.937268 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.937278 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.937291 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.937300 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:17Z","lastTransitionTime":"2025-10-03T12:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:17 crc kubenswrapper[4578]: E1003 12:52:17.948954 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:17Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.951929 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.952010 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.952022 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.952040 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.952054 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:17Z","lastTransitionTime":"2025-10-03T12:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:17 crc kubenswrapper[4578]: E1003 12:52:17.963171 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:17Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.967823 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.967872 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.967884 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.967901 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.967911 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:17Z","lastTransitionTime":"2025-10-03T12:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:17 crc kubenswrapper[4578]: E1003 12:52:17.981462 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:17Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:17 crc kubenswrapper[4578]: E1003 12:52:17.981644 4578 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.983128 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.983158 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.983167 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.983180 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:17 crc kubenswrapper[4578]: I1003 12:52:17.983191 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:17Z","lastTransitionTime":"2025-10-03T12:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.085452 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.085540 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.085551 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.085569 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.085581 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:18Z","lastTransitionTime":"2025-10-03T12:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.188095 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.188135 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.188144 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.188164 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.188173 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:18Z","lastTransitionTime":"2025-10-03T12:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.290448 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.290492 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.290502 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.290516 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.290526 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:18Z","lastTransitionTime":"2025-10-03T12:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.393786 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.393828 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.393838 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.393850 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.393859 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:18Z","lastTransitionTime":"2025-10-03T12:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.496180 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.496219 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.496229 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.496244 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.496255 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:18Z","lastTransitionTime":"2025-10-03T12:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.598216 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.598258 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.598270 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.598286 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.598297 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:18Z","lastTransitionTime":"2025-10-03T12:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.700436 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.700474 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.700494 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.700515 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.700528 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:18Z","lastTransitionTime":"2025-10-03T12:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.802400 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.802450 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.802463 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.802482 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.802495 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:18Z","lastTransitionTime":"2025-10-03T12:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.904442 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.904478 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.904487 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.904500 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.904510 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:18Z","lastTransitionTime":"2025-10-03T12:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.908705 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:18 crc kubenswrapper[4578]: I1003 12:52:18.908776 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:18 crc kubenswrapper[4578]: E1003 12:52:18.908815 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:18 crc kubenswrapper[4578]: E1003 12:52:18.908944 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.006936 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.006976 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.006985 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.006999 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.007008 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:19Z","lastTransitionTime":"2025-10-03T12:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.109605 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.109642 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.109649 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.109662 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.109671 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:19Z","lastTransitionTime":"2025-10-03T12:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.212091 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.212140 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.212156 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.212180 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.212197 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:19Z","lastTransitionTime":"2025-10-03T12:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.315363 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.315429 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.315458 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.315490 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.315527 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:19Z","lastTransitionTime":"2025-10-03T12:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.417648 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.417682 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.417692 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.417706 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.417717 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:19Z","lastTransitionTime":"2025-10-03T12:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.520197 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.520236 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.520244 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.520258 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.520267 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:19Z","lastTransitionTime":"2025-10-03T12:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.622612 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.622756 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.622776 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.622793 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.622803 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:19Z","lastTransitionTime":"2025-10-03T12:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.726038 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.726065 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.726073 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.726085 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.726094 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:19Z","lastTransitionTime":"2025-10-03T12:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.828402 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.828452 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.828466 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.828486 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.828502 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:19Z","lastTransitionTime":"2025-10-03T12:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.908234 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.908306 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:19 crc kubenswrapper[4578]: E1003 12:52:19.908378 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:19 crc kubenswrapper[4578]: E1003 12:52:19.908472 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.930617 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.930662 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.930673 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.930688 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:19 crc kubenswrapper[4578]: I1003 12:52:19.930698 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:19Z","lastTransitionTime":"2025-10-03T12:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.033249 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.033283 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.033291 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.033303 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.033312 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:20Z","lastTransitionTime":"2025-10-03T12:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.135591 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.135657 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.135676 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.135697 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.135707 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:20Z","lastTransitionTime":"2025-10-03T12:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.238379 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.238409 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.238418 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.238450 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.238459 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:20Z","lastTransitionTime":"2025-10-03T12:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.341002 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.341041 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.341054 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.341071 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.341083 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:20Z","lastTransitionTime":"2025-10-03T12:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.443117 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.443176 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.443190 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.443213 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.443223 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:20Z","lastTransitionTime":"2025-10-03T12:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.544898 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.544958 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.544967 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.544980 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.544989 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:20Z","lastTransitionTime":"2025-10-03T12:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.647260 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.647307 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.647318 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.647334 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.647345 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:20Z","lastTransitionTime":"2025-10-03T12:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.748968 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.749008 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.749018 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.749033 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.749045 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:20Z","lastTransitionTime":"2025-10-03T12:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.851784 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.851861 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.851878 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.851893 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.851904 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:20Z","lastTransitionTime":"2025-10-03T12:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.908673 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.908786 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:20 crc kubenswrapper[4578]: E1003 12:52:20.908904 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:20 crc kubenswrapper[4578]: E1003 12:52:20.908980 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.954551 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.954599 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.954609 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.954623 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:20 crc kubenswrapper[4578]: I1003 12:52:20.954646 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:20Z","lastTransitionTime":"2025-10-03T12:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.057105 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.057167 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.057180 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.057216 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.057230 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:21Z","lastTransitionTime":"2025-10-03T12:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.159871 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.159936 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.159950 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.159972 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.159985 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:21Z","lastTransitionTime":"2025-10-03T12:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.262669 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.262702 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.262710 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.262722 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.262732 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:21Z","lastTransitionTime":"2025-10-03T12:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.365356 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.365409 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.365418 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.365432 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.365441 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:21Z","lastTransitionTime":"2025-10-03T12:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.468206 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.468246 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.468259 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.468273 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.468282 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:21Z","lastTransitionTime":"2025-10-03T12:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.570944 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.571003 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.571012 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.571025 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.571036 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:21Z","lastTransitionTime":"2025-10-03T12:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.673591 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.673661 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.673671 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.673685 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.673696 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:21Z","lastTransitionTime":"2025-10-03T12:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.776738 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.776786 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.776798 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.776814 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.776826 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:21Z","lastTransitionTime":"2025-10-03T12:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.879244 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.879276 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.879290 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.879303 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.879312 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:21Z","lastTransitionTime":"2025-10-03T12:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.908606 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.908694 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:21 crc kubenswrapper[4578]: E1003 12:52:21.908800 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:21 crc kubenswrapper[4578]: E1003 12:52:21.908902 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.981844 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.981887 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.981901 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.981918 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:21 crc kubenswrapper[4578]: I1003 12:52:21.981935 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:21Z","lastTransitionTime":"2025-10-03T12:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.084580 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.084660 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.084676 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.084691 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.084706 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:22Z","lastTransitionTime":"2025-10-03T12:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.186307 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.186346 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.186359 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.186374 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.186386 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:22Z","lastTransitionTime":"2025-10-03T12:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.289131 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.289220 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.289233 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.289248 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.289258 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:22Z","lastTransitionTime":"2025-10-03T12:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.392457 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-c4jgn_fed3968f-4bb0-4209-a377-16c26f9ea070/kube-multus/0.log" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.392954 4578 generic.go:334] "Generic (PLEG): container finished" podID="fed3968f-4bb0-4209-a377-16c26f9ea070" containerID="7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3" exitCode=1 Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.393032 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-c4jgn" event={"ID":"fed3968f-4bb0-4209-a377-16c26f9ea070","Type":"ContainerDied","Data":"7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3"} Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.393248 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.393276 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.393285 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.393296 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.393305 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:22Z","lastTransitionTime":"2025-10-03T12:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.393754 4578 scope.go:117] "RemoveContainer" containerID="7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.410140 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:22Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.432236 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:00Z\\\",\\\"message\\\":\\\"rvices.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:52:00.501792 6175 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nF1003 12:52:00.501828 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z]\\\\nI1003 12:52:00.501838 6175 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nI1003 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:22Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.443058 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:22Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.455134 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b4650253599f9031287cc43642dff608baa3920e6dce60b433a7c642cece88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:22Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.469666 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:22Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.485895 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:22Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.495509 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:22Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.495817 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.495851 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.495861 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.495877 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.495888 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:22Z","lastTransitionTime":"2025-10-03T12:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.506404 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:21Z\\\",\\\"message\\\":\\\"2025-10-03T12:51:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_57184fea-0a56-4895-9b4d-dbeabc7e5d55\\\\n2025-10-03T12:51:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_57184fea-0a56-4895-9b4d-dbeabc7e5d55 to /host/opt/cni/bin/\\\\n2025-10-03T12:51:36Z [verbose] multus-daemon started\\\\n2025-10-03T12:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-03T12:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:22Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.515432 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:22Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.525780 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:22Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.537223 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a524da53-3b46-423d-97ae-0e7d131b958f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d4f0ede4a9f9262ffe08a7abc4366faf1001b25f3a6a35dfac42c3d8362030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ba8c08b2dceaf712d8d6b27f5c3a4b6fd3a204fc494dfb8ec33ed7ce8bc16e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1fb8c53aa23e2c70af59f270c04d09599d79718098c17a809ace5f72797303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:22Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.550371 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:22Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.562170 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:22Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.572419 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:22Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.583151 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:22Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.596424 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:22Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.597881 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.597912 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.597924 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.597940 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.597952 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:22Z","lastTransitionTime":"2025-10-03T12:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.700296 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.700338 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.700349 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.700364 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.700375 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:22Z","lastTransitionTime":"2025-10-03T12:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.802262 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.802325 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.802336 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.802356 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.802369 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:22Z","lastTransitionTime":"2025-10-03T12:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.904721 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.904769 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.904778 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.904791 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.904801 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:22Z","lastTransitionTime":"2025-10-03T12:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.909044 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:22 crc kubenswrapper[4578]: E1003 12:52:22.909130 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:22 crc kubenswrapper[4578]: I1003 12:52:22.909045 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:22 crc kubenswrapper[4578]: E1003 12:52:22.909195 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.006712 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.006758 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.006772 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.006789 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.006802 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:23Z","lastTransitionTime":"2025-10-03T12:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.108329 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.108366 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.108377 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.108397 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.108410 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:23Z","lastTransitionTime":"2025-10-03T12:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.210447 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.210479 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.210489 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.210502 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.210510 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:23Z","lastTransitionTime":"2025-10-03T12:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.313051 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.313097 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.313111 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.313133 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.313149 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:23Z","lastTransitionTime":"2025-10-03T12:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.397223 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-c4jgn_fed3968f-4bb0-4209-a377-16c26f9ea070/kube-multus/0.log" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.397274 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-c4jgn" event={"ID":"fed3968f-4bb0-4209-a377-16c26f9ea070","Type":"ContainerStarted","Data":"650520feb6cfcf5c6aea04b4a990593eab6f2a2ddec00e2270105f25ee9a067d"} Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.409012 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a524da53-3b46-423d-97ae-0e7d131b958f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d4f0ede4a9f9262ffe08a7abc4366faf1001b25f3a6a35dfac42c3d8362030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ba8c08b2dceaf712d8d6b27f5c3a4b6fd3a204fc494dfb8ec33ed7ce8bc16e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1fb8c53aa23e2c70af59f270c04d09599d79718098c17a809ace5f72797303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:23Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.415952 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.415986 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.415996 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.416011 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.416020 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:23Z","lastTransitionTime":"2025-10-03T12:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.421478 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:23Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.431558 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:23Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.440325 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:23Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.450024 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:23Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.459394 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:23Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.469128 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:23Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.481239 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:23Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.491934 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b4650253599f9031287cc43642dff608baa3920e6dce60b433a7c642cece88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:23Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.506251 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:23Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.516316 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:23Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.517829 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.517850 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.517858 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.517871 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.517883 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:23Z","lastTransitionTime":"2025-10-03T12:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.533684 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:00Z\\\",\\\"message\\\":\\\"rvices.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:52:00.501792 6175 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nF1003 12:52:00.501828 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z]\\\\nI1003 12:52:00.501838 6175 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nI1003 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:23Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.544185 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:23Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.554571 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650520feb6cfcf5c6aea04b4a990593eab6f2a2ddec00e2270105f25ee9a067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:21Z\\\",\\\"message\\\":\\\"2025-10-03T12:51:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_57184fea-0a56-4895-9b4d-dbeabc7e5d55\\\\n2025-10-03T12:51:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_57184fea-0a56-4895-9b4d-dbeabc7e5d55 to /host/opt/cni/bin/\\\\n2025-10-03T12:51:36Z [verbose] multus-daemon started\\\\n2025-10-03T12:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-03T12:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:23Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.563671 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:23Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.572122 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:23Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.619775 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.619830 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.619843 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.619882 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.619894 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:23Z","lastTransitionTime":"2025-10-03T12:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.722131 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.722281 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.722296 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.722313 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.722324 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:23Z","lastTransitionTime":"2025-10-03T12:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.824676 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.824702 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.824709 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.824722 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.824731 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:23Z","lastTransitionTime":"2025-10-03T12:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.908096 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:23 crc kubenswrapper[4578]: E1003 12:52:23.908266 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.908536 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:23 crc kubenswrapper[4578]: E1003 12:52:23.908607 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.927258 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.927298 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.927310 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.927326 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:23 crc kubenswrapper[4578]: I1003 12:52:23.927339 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:23Z","lastTransitionTime":"2025-10-03T12:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.029835 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.029890 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.029903 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.029925 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.029937 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:24Z","lastTransitionTime":"2025-10-03T12:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.132000 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.132053 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.132065 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.132086 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.132105 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:24Z","lastTransitionTime":"2025-10-03T12:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.234355 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.234396 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.234406 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.234434 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.234445 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:24Z","lastTransitionTime":"2025-10-03T12:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.337056 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.337095 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.337107 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.337127 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.337140 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:24Z","lastTransitionTime":"2025-10-03T12:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.439180 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.439222 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.439234 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.439250 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.439262 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:24Z","lastTransitionTime":"2025-10-03T12:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.541257 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.541294 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.541304 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.541319 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.541329 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:24Z","lastTransitionTime":"2025-10-03T12:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.644054 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.644094 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.644102 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.644115 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.644125 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:24Z","lastTransitionTime":"2025-10-03T12:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.745977 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.746014 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.746023 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.746038 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.746047 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:24Z","lastTransitionTime":"2025-10-03T12:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.848183 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.848234 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.848243 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.848257 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.848266 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:24Z","lastTransitionTime":"2025-10-03T12:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.908900 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.908963 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:24 crc kubenswrapper[4578]: E1003 12:52:24.909044 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:24 crc kubenswrapper[4578]: E1003 12:52:24.909120 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.920871 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:24Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.938424 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:00Z\\\",\\\"message\\\":\\\"rvices.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:52:00.501792 6175 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nF1003 12:52:00.501828 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z]\\\\nI1003 12:52:00.501838 6175 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nI1003 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:24Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.949837 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:24Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.950139 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.950171 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.950179 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.950192 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.950200 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:24Z","lastTransitionTime":"2025-10-03T12:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.961989 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b4650253599f9031287cc43642dff608baa3920e6dce60b433a7c642cece88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:24Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.974037 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:24Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.985894 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:24Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:24 crc kubenswrapper[4578]: I1003 12:52:24.995672 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:24Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.009695 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650520feb6cfcf5c6aea04b4a990593eab6f2a2ddec00e2270105f25ee9a067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:21Z\\\",\\\"message\\\":\\\"2025-10-03T12:51:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_57184fea-0a56-4895-9b4d-dbeabc7e5d55\\\\n2025-10-03T12:51:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_57184fea-0a56-4895-9b4d-dbeabc7e5d55 to /host/opt/cni/bin/\\\\n2025-10-03T12:51:36Z [verbose] multus-daemon started\\\\n2025-10-03T12:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-03T12:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:25Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.022308 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:25Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.035678 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:25Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.047086 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a524da53-3b46-423d-97ae-0e7d131b958f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d4f0ede4a9f9262ffe08a7abc4366faf1001b25f3a6a35dfac42c3d8362030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ba8c08b2dceaf712d8d6b27f5c3a4b6fd3a204fc494dfb8ec33ed7ce8bc16e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1fb8c53aa23e2c70af59f270c04d09599d79718098c17a809ace5f72797303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:25Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.054284 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.054339 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.054348 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.054361 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.054370 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:25Z","lastTransitionTime":"2025-10-03T12:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.058502 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:25Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.069861 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:25Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.079361 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:25Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.089484 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:25Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.103208 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:25Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.156552 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.156595 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.156607 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.156623 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.156652 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:25Z","lastTransitionTime":"2025-10-03T12:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.259153 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.259197 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.259206 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.259249 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.259259 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:25Z","lastTransitionTime":"2025-10-03T12:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.361531 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.361584 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.361595 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.361609 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.361619 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:25Z","lastTransitionTime":"2025-10-03T12:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.464192 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.464257 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.464272 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.464289 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.464302 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:25Z","lastTransitionTime":"2025-10-03T12:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.566664 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.566746 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.566756 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.566771 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.566780 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:25Z","lastTransitionTime":"2025-10-03T12:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.668792 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.668827 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.668852 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.668868 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.668879 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:25Z","lastTransitionTime":"2025-10-03T12:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.771531 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.771566 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.771577 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.771636 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.771650 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:25Z","lastTransitionTime":"2025-10-03T12:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.874020 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.874084 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.874096 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.874114 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.874126 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:25Z","lastTransitionTime":"2025-10-03T12:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.909068 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:25 crc kubenswrapper[4578]: E1003 12:52:25.909191 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.909355 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:25 crc kubenswrapper[4578]: E1003 12:52:25.909401 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.977097 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.977169 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.977180 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.977194 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:25 crc kubenswrapper[4578]: I1003 12:52:25.977203 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:25Z","lastTransitionTime":"2025-10-03T12:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.079733 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.079772 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.079781 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.079794 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.079803 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:26Z","lastTransitionTime":"2025-10-03T12:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.182008 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.182052 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.182070 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.182091 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.182108 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:26Z","lastTransitionTime":"2025-10-03T12:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.284778 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.284821 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.284830 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.284844 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.284854 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:26Z","lastTransitionTime":"2025-10-03T12:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.386455 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.386493 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.386504 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.386519 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.386531 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:26Z","lastTransitionTime":"2025-10-03T12:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.488535 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.488562 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.488569 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.488583 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.488591 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:26Z","lastTransitionTime":"2025-10-03T12:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.591701 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.591750 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.591760 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.591775 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.591785 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:26Z","lastTransitionTime":"2025-10-03T12:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.694452 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.694502 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.694515 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.694531 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.694543 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:26Z","lastTransitionTime":"2025-10-03T12:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.796388 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.796442 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.796450 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.796464 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.796474 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:26Z","lastTransitionTime":"2025-10-03T12:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.898054 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.898104 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.898116 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.898134 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.898147 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:26Z","lastTransitionTime":"2025-10-03T12:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.908318 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:26 crc kubenswrapper[4578]: I1003 12:52:26.908378 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:26 crc kubenswrapper[4578]: E1003 12:52:26.908419 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:26 crc kubenswrapper[4578]: E1003 12:52:26.908510 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.003356 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.003400 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.003409 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.003424 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.003433 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:27Z","lastTransitionTime":"2025-10-03T12:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.105774 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.105817 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.105825 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.105839 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.105850 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:27Z","lastTransitionTime":"2025-10-03T12:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.208174 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.208219 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.208232 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.208249 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.208262 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:27Z","lastTransitionTime":"2025-10-03T12:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.311265 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.311304 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.311315 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.311332 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.311344 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:27Z","lastTransitionTime":"2025-10-03T12:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.413089 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.413727 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.413755 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.413769 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.413780 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:27Z","lastTransitionTime":"2025-10-03T12:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.515727 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.515770 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.515784 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.515800 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.515810 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:27Z","lastTransitionTime":"2025-10-03T12:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.618464 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.618507 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.618518 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.618537 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.618549 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:27Z","lastTransitionTime":"2025-10-03T12:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.720819 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.720859 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.720868 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.720883 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.720891 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:27Z","lastTransitionTime":"2025-10-03T12:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.822858 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.822894 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.822904 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.822918 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.822928 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:27Z","lastTransitionTime":"2025-10-03T12:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.908172 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:27 crc kubenswrapper[4578]: E1003 12:52:27.908282 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.908172 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:27 crc kubenswrapper[4578]: E1003 12:52:27.908470 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.909045 4578 scope.go:117] "RemoveContainer" containerID="578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.919230 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.925310 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.925351 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.925360 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.925375 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:27 crc kubenswrapper[4578]: I1003 12:52:27.925386 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:27Z","lastTransitionTime":"2025-10-03T12:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.027247 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.027285 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.027297 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.027313 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.027324 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:28Z","lastTransitionTime":"2025-10-03T12:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.121157 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.121495 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.121503 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.121518 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.121527 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:28Z","lastTransitionTime":"2025-10-03T12:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:28 crc kubenswrapper[4578]: E1003 12:52:28.133513 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.136947 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.136972 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.136983 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.136999 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.137011 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:28Z","lastTransitionTime":"2025-10-03T12:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:28 crc kubenswrapper[4578]: E1003 12:52:28.149225 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.152600 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.152653 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.152665 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.152681 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.152691 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:28Z","lastTransitionTime":"2025-10-03T12:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:28 crc kubenswrapper[4578]: E1003 12:52:28.165353 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.169462 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.169505 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.169516 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.169534 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.169547 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:28Z","lastTransitionTime":"2025-10-03T12:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:28 crc kubenswrapper[4578]: E1003 12:52:28.181772 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.186827 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.186870 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.186880 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.186897 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.186908 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:28Z","lastTransitionTime":"2025-10-03T12:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:28 crc kubenswrapper[4578]: E1003 12:52:28.203083 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: E1003 12:52:28.203251 4578 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.205453 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.205490 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.205502 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.205517 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.205528 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:28Z","lastTransitionTime":"2025-10-03T12:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.308097 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.308135 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.308145 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.308159 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.308170 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:28Z","lastTransitionTime":"2025-10-03T12:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.410281 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.410319 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.410328 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.410344 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.410354 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:28Z","lastTransitionTime":"2025-10-03T12:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.412466 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8qvh9_ebaa96c5-be7c-4f05-819d-f37e1ceddac8/ovnkube-controller/2.log" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.415668 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerStarted","Data":"508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c"} Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.416679 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.430335 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.445165 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b4650253599f9031287cc43642dff608baa3920e6dce60b433a7c642cece88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.464867 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.493703 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.514068 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.514135 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.514146 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.514179 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.514194 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:28Z","lastTransitionTime":"2025-10-03T12:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.517087 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:00Z\\\",\\\"message\\\":\\\"rvices.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:52:00.501792 6175 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nF1003 12:52:00.501828 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z]\\\\nI1003 12:52:00.501838 6175 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nI1003 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.530781 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e43f233-de60-499f-bcc6-8e782a42f31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ede2cac008ac6d78a301b36ea4cdd25ccc11646c529d397c230d264590fcbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60be5924c0b93bbd78ef07fc29492fd17b85c06407edcecfb7763032326000b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60be5924c0b93bbd78ef07fc29492fd17b85c06407edcecfb7763032326000b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.550677 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650520feb6cfcf5c6aea04b4a990593eab6f2a2ddec00e2270105f25ee9a067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:21Z\\\",\\\"message\\\":\\\"2025-10-03T12:51:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_57184fea-0a56-4895-9b4d-dbeabc7e5d55\\\\n2025-10-03T12:51:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_57184fea-0a56-4895-9b4d-dbeabc7e5d55 to /host/opt/cni/bin/\\\\n2025-10-03T12:51:36Z [verbose] multus-daemon started\\\\n2025-10-03T12:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-03T12:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.564621 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.582260 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.596555 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a524da53-3b46-423d-97ae-0e7d131b958f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d4f0ede4a9f9262ffe08a7abc4366faf1001b25f3a6a35dfac42c3d8362030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ba8c08b2dceaf712d8d6b27f5c3a4b6fd3a204fc494dfb8ec33ed7ce8bc16e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1fb8c53aa23e2c70af59f270c04d09599d79718098c17a809ace5f72797303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.613498 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.616998 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.617037 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.617046 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.617058 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.617067 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:28Z","lastTransitionTime":"2025-10-03T12:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.632145 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.648152 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.663336 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.677564 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.693045 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.707022 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:28Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.719791 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.719817 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.719826 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.719838 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.719847 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:28Z","lastTransitionTime":"2025-10-03T12:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.822731 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.822788 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.822798 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.822817 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.822829 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:28Z","lastTransitionTime":"2025-10-03T12:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.908797 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.908854 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:28 crc kubenswrapper[4578]: E1003 12:52:28.909324 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:28 crc kubenswrapper[4578]: E1003 12:52:28.909344 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.926268 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.926330 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.926347 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.926370 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:28 crc kubenswrapper[4578]: I1003 12:52:28.926385 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:28Z","lastTransitionTime":"2025-10-03T12:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.029850 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.029892 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.029902 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.029927 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.029937 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:29Z","lastTransitionTime":"2025-10-03T12:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.132361 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.132688 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.132849 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.132950 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.133047 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:29Z","lastTransitionTime":"2025-10-03T12:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.236823 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.236886 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.236904 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.236929 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.236947 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:29Z","lastTransitionTime":"2025-10-03T12:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.338908 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.339159 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.339238 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.339297 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.339382 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:29Z","lastTransitionTime":"2025-10-03T12:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.419831 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8qvh9_ebaa96c5-be7c-4f05-819d-f37e1ceddac8/ovnkube-controller/3.log" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.421044 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8qvh9_ebaa96c5-be7c-4f05-819d-f37e1ceddac8/ovnkube-controller/2.log" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.423455 4578 generic.go:334] "Generic (PLEG): container finished" podID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerID="508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c" exitCode=1 Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.423488 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerDied","Data":"508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c"} Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.423518 4578 scope.go:117] "RemoveContainer" containerID="578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.424095 4578 scope.go:117] "RemoveContainer" containerID="508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c" Oct 03 12:52:29 crc kubenswrapper[4578]: E1003 12:52:29.424222 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.437080 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e43f233-de60-499f-bcc6-8e782a42f31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ede2cac008ac6d78a301b36ea4cdd25ccc11646c529d397c230d264590fcbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60be5924c0b93bbd78ef07fc29492fd17b85c06407edcecfb7763032326000b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60be5924c0b93bbd78ef07fc29492fd17b85c06407edcecfb7763032326000b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:29Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.441052 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.441088 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.441100 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.441116 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.441127 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:29Z","lastTransitionTime":"2025-10-03T12:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.448874 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650520feb6cfcf5c6aea04b4a990593eab6f2a2ddec00e2270105f25ee9a067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:21Z\\\",\\\"message\\\":\\\"2025-10-03T12:51:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_57184fea-0a56-4895-9b4d-dbeabc7e5d55\\\\n2025-10-03T12:51:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_57184fea-0a56-4895-9b4d-dbeabc7e5d55 to /host/opt/cni/bin/\\\\n2025-10-03T12:51:36Z [verbose] multus-daemon started\\\\n2025-10-03T12:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-03T12:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:29Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.458748 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:29Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.467375 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:29Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.476906 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a524da53-3b46-423d-97ae-0e7d131b958f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d4f0ede4a9f9262ffe08a7abc4366faf1001b25f3a6a35dfac42c3d8362030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ba8c08b2dceaf712d8d6b27f5c3a4b6fd3a204fc494dfb8ec33ed7ce8bc16e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1fb8c53aa23e2c70af59f270c04d09599d79718098c17a809ace5f72797303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:29Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.487596 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:29Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.498017 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:29Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.507434 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:29Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.516818 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:29Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.527214 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:29Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.537244 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:29Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.543568 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.543598 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.543608 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.543622 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.543653 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:29Z","lastTransitionTime":"2025-10-03T12:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.551419 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:29Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.564794 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b4650253599f9031287cc43642dff608baa3920e6dce60b433a7c642cece88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:29Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.577591 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:29Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.590373 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:29Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.610222 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://578ba36fed8fe753df2c0168c89b7c27b268edb78acc1a27ea66761cde8672ce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:00Z\\\",\\\"message\\\":\\\"rvices.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:52:00.501792 6175 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nF1003 12:52:00.501828 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:00Z is after 2025-08-24T17:21:41Z]\\\\nI1003 12:52:00.501838 6175 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nI1003 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:29Z\\\",\\\"message\\\":\\\"0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:52:29.037364 6557 services_controller.go:444] Built service openshift-machine-config-operator/machine-config-controller LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1003 12:52:29.037373 6557 services_controller.go:445] Built service openshift-machine-config-operator/machine-config-controller LB template configs for network=default: []services.lbConfig(nil)\\\\nI1003 12:52:29.037391 6557 services_controller.go:451] Built service openshift-machine-config-operator/machine-config-controller cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.16\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Router\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:52:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:29Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.619431 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:29Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.646353 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.646388 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.646398 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.646411 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.646420 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:29Z","lastTransitionTime":"2025-10-03T12:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.749077 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.749118 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.749132 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.749148 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.749170 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:29Z","lastTransitionTime":"2025-10-03T12:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.851265 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.851359 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.851379 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.851408 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.851427 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:29Z","lastTransitionTime":"2025-10-03T12:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.908276 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.908348 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:29 crc kubenswrapper[4578]: E1003 12:52:29.908480 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:29 crc kubenswrapper[4578]: E1003 12:52:29.908604 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.954276 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.954720 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.954790 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.954893 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:29 crc kubenswrapper[4578]: I1003 12:52:29.954997 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:29Z","lastTransitionTime":"2025-10-03T12:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.057896 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.058174 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.058241 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.058300 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.058358 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:30Z","lastTransitionTime":"2025-10-03T12:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.160867 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.161236 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.161399 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.161485 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.161567 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:30Z","lastTransitionTime":"2025-10-03T12:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.264723 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.265059 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.265177 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.265258 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.265324 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:30Z","lastTransitionTime":"2025-10-03T12:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.368008 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.368044 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.368055 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.368070 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.368082 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:30Z","lastTransitionTime":"2025-10-03T12:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.429249 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8qvh9_ebaa96c5-be7c-4f05-819d-f37e1ceddac8/ovnkube-controller/3.log" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.434416 4578 scope.go:117] "RemoveContainer" containerID="508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c" Oct 03 12:52:30 crc kubenswrapper[4578]: E1003 12:52:30.434575 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.449153 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:30Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.462136 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:30Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.471021 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.471076 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.471088 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.471105 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.471116 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:30Z","lastTransitionTime":"2025-10-03T12:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.482710 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:29Z\\\",\\\"message\\\":\\\"0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:52:29.037364 6557 services_controller.go:444] Built service openshift-machine-config-operator/machine-config-controller LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1003 12:52:29.037373 6557 services_controller.go:445] Built service openshift-machine-config-operator/machine-config-controller LB template configs for network=default: []services.lbConfig(nil)\\\\nI1003 12:52:29.037391 6557 services_controller.go:451] Built service openshift-machine-config-operator/machine-config-controller cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.16\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Router\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:52:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:30Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.496392 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:30Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.517522 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b4650253599f9031287cc43642dff608baa3920e6dce60b433a7c642cece88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:30Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.532857 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650520feb6cfcf5c6aea04b4a990593eab6f2a2ddec00e2270105f25ee9a067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:21Z\\\",\\\"message\\\":\\\"2025-10-03T12:51:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_57184fea-0a56-4895-9b4d-dbeabc7e5d55\\\\n2025-10-03T12:51:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_57184fea-0a56-4895-9b4d-dbeabc7e5d55 to /host/opt/cni/bin/\\\\n2025-10-03T12:51:36Z [verbose] multus-daemon started\\\\n2025-10-03T12:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-03T12:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:30Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.546590 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:30Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.560189 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:30Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.571550 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e43f233-de60-499f-bcc6-8e782a42f31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ede2cac008ac6d78a301b36ea4cdd25ccc11646c529d397c230d264590fcbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60be5924c0b93bbd78ef07fc29492fd17b85c06407edcecfb7763032326000b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60be5924c0b93bbd78ef07fc29492fd17b85c06407edcecfb7763032326000b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:30Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.573480 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.573527 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.573539 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.573552 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.573561 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:30Z","lastTransitionTime":"2025-10-03T12:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.585586 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:30Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.600330 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:30Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.612791 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:30Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.624499 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a524da53-3b46-423d-97ae-0e7d131b958f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d4f0ede4a9f9262ffe08a7abc4366faf1001b25f3a6a35dfac42c3d8362030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ba8c08b2dceaf712d8d6b27f5c3a4b6fd3a204fc494dfb8ec33ed7ce8bc16e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1fb8c53aa23e2c70af59f270c04d09599d79718098c17a809ace5f72797303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:30Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.637123 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:30Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.656171 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:30Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.670065 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:30Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.676841 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.676921 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.676998 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.677049 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.677069 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:30Z","lastTransitionTime":"2025-10-03T12:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.689190 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:30Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.780282 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.780321 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.780332 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.780348 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.780357 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:30Z","lastTransitionTime":"2025-10-03T12:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.882801 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.882856 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.882867 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.882881 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.882890 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:30Z","lastTransitionTime":"2025-10-03T12:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.908698 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:30 crc kubenswrapper[4578]: E1003 12:52:30.908891 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.908700 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:30 crc kubenswrapper[4578]: E1003 12:52:30.909292 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.985365 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.985423 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.985431 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.985443 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:30 crc kubenswrapper[4578]: I1003 12:52:30.985451 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:30Z","lastTransitionTime":"2025-10-03T12:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.088269 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.088318 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.088330 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.088355 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.088372 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:31Z","lastTransitionTime":"2025-10-03T12:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.190844 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.190887 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.190897 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.190909 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.190918 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:31Z","lastTransitionTime":"2025-10-03T12:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.293172 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.293203 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.293212 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.293225 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.293234 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:31Z","lastTransitionTime":"2025-10-03T12:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.395452 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.395721 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.395804 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.395934 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.395997 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:31Z","lastTransitionTime":"2025-10-03T12:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.497720 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.497752 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.497762 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.497773 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.497785 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:31Z","lastTransitionTime":"2025-10-03T12:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.601695 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.602225 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.602441 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.602989 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.603236 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:31Z","lastTransitionTime":"2025-10-03T12:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.706300 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.706330 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.706341 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.706356 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.706366 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:31Z","lastTransitionTime":"2025-10-03T12:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.808954 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.809203 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.809270 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.809329 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.809397 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:31Z","lastTransitionTime":"2025-10-03T12:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.908569 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:31 crc kubenswrapper[4578]: E1003 12:52:31.908707 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.908569 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:31 crc kubenswrapper[4578]: E1003 12:52:31.908919 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.912116 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.912163 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.912190 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.912207 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:31 crc kubenswrapper[4578]: I1003 12:52:31.912217 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:31Z","lastTransitionTime":"2025-10-03T12:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.014204 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.014455 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.014589 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.014700 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.014788 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:32Z","lastTransitionTime":"2025-10-03T12:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.117445 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.117494 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.117502 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.117514 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.117523 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:32Z","lastTransitionTime":"2025-10-03T12:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.220000 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.220038 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.220046 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.220061 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.220070 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:32Z","lastTransitionTime":"2025-10-03T12:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.322448 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.322478 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.322488 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.322501 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.322511 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:32Z","lastTransitionTime":"2025-10-03T12:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.425086 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.425374 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.425473 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.425557 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.425615 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:32Z","lastTransitionTime":"2025-10-03T12:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.527903 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.528417 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.528515 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.528609 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.528724 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:32Z","lastTransitionTime":"2025-10-03T12:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.631318 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.631377 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.631389 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.631404 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.631415 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:32Z","lastTransitionTime":"2025-10-03T12:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.734179 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.734599 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.734784 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.734970 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.735134 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:32Z","lastTransitionTime":"2025-10-03T12:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.837346 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.837380 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.837391 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.837408 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.837418 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:32Z","lastTransitionTime":"2025-10-03T12:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.908823 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.908842 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:32 crc kubenswrapper[4578]: E1003 12:52:32.909366 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:32 crc kubenswrapper[4578]: E1003 12:52:32.909561 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.940270 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.940307 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.940317 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.940333 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:32 crc kubenswrapper[4578]: I1003 12:52:32.940344 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:32Z","lastTransitionTime":"2025-10-03T12:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.043369 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.043423 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.043436 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.043458 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.043474 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:33Z","lastTransitionTime":"2025-10-03T12:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.145814 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.145888 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.145897 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.145912 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.145922 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:33Z","lastTransitionTime":"2025-10-03T12:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.248149 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.248185 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.248194 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.248207 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.248219 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:33Z","lastTransitionTime":"2025-10-03T12:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.350420 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.350459 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.350468 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.350489 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.350499 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:33Z","lastTransitionTime":"2025-10-03T12:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.452186 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.452218 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.452226 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.452239 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.452247 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:33Z","lastTransitionTime":"2025-10-03T12:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.554180 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.554212 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.554219 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.554232 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.554241 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:33Z","lastTransitionTime":"2025-10-03T12:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.656815 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.656851 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.656861 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.656873 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.656883 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:33Z","lastTransitionTime":"2025-10-03T12:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.759361 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.759393 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.759401 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.759413 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.759422 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:33Z","lastTransitionTime":"2025-10-03T12:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.861419 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.861451 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.861462 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.861476 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.861486 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:33Z","lastTransitionTime":"2025-10-03T12:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.907987 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.907992 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:33 crc kubenswrapper[4578]: E1003 12:52:33.908104 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:33 crc kubenswrapper[4578]: E1003 12:52:33.908204 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.963802 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.963839 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.963847 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.963860 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:33 crc kubenswrapper[4578]: I1003 12:52:33.963868 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:33Z","lastTransitionTime":"2025-10-03T12:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.066380 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.066416 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.066427 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.066442 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.066456 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:34Z","lastTransitionTime":"2025-10-03T12:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.168783 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.168843 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.168852 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.168866 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.168878 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:34Z","lastTransitionTime":"2025-10-03T12:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.271318 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.271358 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.271367 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.271384 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.271395 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:34Z","lastTransitionTime":"2025-10-03T12:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.373317 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.373360 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.373373 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.373389 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.373399 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:34Z","lastTransitionTime":"2025-10-03T12:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.475522 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.475558 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.475570 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.475588 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.475599 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:34Z","lastTransitionTime":"2025-10-03T12:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.577987 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.578031 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.578043 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.578059 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.578070 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:34Z","lastTransitionTime":"2025-10-03T12:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.680171 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.680202 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.680209 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.680222 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.680232 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:34Z","lastTransitionTime":"2025-10-03T12:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.782656 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.782688 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.782699 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.782714 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.782731 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:34Z","lastTransitionTime":"2025-10-03T12:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.884759 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.884789 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.884797 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.884810 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.884818 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:34Z","lastTransitionTime":"2025-10-03T12:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.908581 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.908645 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:34 crc kubenswrapper[4578]: E1003 12:52:34.908707 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:34 crc kubenswrapper[4578]: E1003 12:52:34.908833 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.927085 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:29Z\\\",\\\"message\\\":\\\"0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:52:29.037364 6557 services_controller.go:444] Built service openshift-machine-config-operator/machine-config-controller LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1003 12:52:29.037373 6557 services_controller.go:445] Built service openshift-machine-config-operator/machine-config-controller LB template configs for network=default: []services.lbConfig(nil)\\\\nI1003 12:52:29.037391 6557 services_controller.go:451] Built service openshift-machine-config-operator/machine-config-controller cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.16\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Router\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:52:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.938414 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.952304 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b4650253599f9031287cc43642dff608baa3920e6dce60b433a7c642cece88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.967900 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.980234 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.987677 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.987730 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.987741 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.987758 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.987769 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:34Z","lastTransitionTime":"2025-10-03T12:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.989382 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:34 crc kubenswrapper[4578]: I1003 12:52:34.998324 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e43f233-de60-499f-bcc6-8e782a42f31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ede2cac008ac6d78a301b36ea4cdd25ccc11646c529d397c230d264590fcbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60be5924c0b93bbd78ef07fc29492fd17b85c06407edcecfb7763032326000b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60be5924c0b93bbd78ef07fc29492fd17b85c06407edcecfb7763032326000b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:34Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.015791 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650520feb6cfcf5c6aea04b4a990593eab6f2a2ddec00e2270105f25ee9a067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:21Z\\\",\\\"message\\\":\\\"2025-10-03T12:51:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_57184fea-0a56-4895-9b4d-dbeabc7e5d55\\\\n2025-10-03T12:51:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_57184fea-0a56-4895-9b4d-dbeabc7e5d55 to /host/opt/cni/bin/\\\\n2025-10-03T12:51:36Z [verbose] multus-daemon started\\\\n2025-10-03T12:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-03T12:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.027514 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.040304 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.050789 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a524da53-3b46-423d-97ae-0e7d131b958f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d4f0ede4a9f9262ffe08a7abc4366faf1001b25f3a6a35dfac42c3d8362030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ba8c08b2dceaf712d8d6b27f5c3a4b6fd3a204fc494dfb8ec33ed7ce8bc16e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1fb8c53aa23e2c70af59f270c04d09599d79718098c17a809ace5f72797303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.061486 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.071359 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.079740 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.090919 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.091176 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.091238 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.091251 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.091308 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.091323 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:35Z","lastTransitionTime":"2025-10-03T12:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.102253 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.114277 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:35Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.193444 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.193482 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.193497 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.193515 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.193526 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:35Z","lastTransitionTime":"2025-10-03T12:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.296124 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.296159 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.296172 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.296187 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.296196 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:35Z","lastTransitionTime":"2025-10-03T12:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.398326 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.398371 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.398383 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.398400 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.398414 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:35Z","lastTransitionTime":"2025-10-03T12:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.500488 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.500527 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.500537 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.500553 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.500563 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:35Z","lastTransitionTime":"2025-10-03T12:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.603151 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.603222 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.603238 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.603256 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.603268 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:35Z","lastTransitionTime":"2025-10-03T12:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.705642 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.705669 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.705677 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.705689 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.705698 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:35Z","lastTransitionTime":"2025-10-03T12:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.746599 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:52:35 crc kubenswrapper[4578]: E1003 12:52:35.746802 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:39.74677173 +0000 UTC m=+155.545243924 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.808366 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.808407 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.808417 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.808453 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.808465 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:35Z","lastTransitionTime":"2025-10-03T12:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.847657 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.847731 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.847794 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:35 crc kubenswrapper[4578]: E1003 12:52:35.847858 4578 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 12:52:35 crc kubenswrapper[4578]: E1003 12:52:35.847889 4578 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 12:52:35 crc kubenswrapper[4578]: E1003 12:52:35.847936 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 12:53:39.847917945 +0000 UTC m=+155.646390129 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 03 12:52:35 crc kubenswrapper[4578]: E1003 12:52:35.847953 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-03 12:53:39.847945946 +0000 UTC m=+155.646418130 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 03 12:52:35 crc kubenswrapper[4578]: E1003 12:52:35.848085 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 12:52:35 crc kubenswrapper[4578]: E1003 12:52:35.848117 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 12:52:35 crc kubenswrapper[4578]: E1003 12:52:35.848130 4578 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:52:35 crc kubenswrapper[4578]: E1003 12:52:35.848184 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-03 12:53:39.848167763 +0000 UTC m=+155.646639947 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.848219 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:35 crc kubenswrapper[4578]: E1003 12:52:35.848334 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 03 12:52:35 crc kubenswrapper[4578]: E1003 12:52:35.848345 4578 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 03 12:52:35 crc kubenswrapper[4578]: E1003 12:52:35.848352 4578 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:52:35 crc kubenswrapper[4578]: E1003 12:52:35.848386 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-03 12:53:39.84837944 +0000 UTC m=+155.646851624 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.908953 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:35 crc kubenswrapper[4578]: E1003 12:52:35.909084 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.909311 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:35 crc kubenswrapper[4578]: E1003 12:52:35.909364 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.910827 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.910848 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.910869 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.910882 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:35 crc kubenswrapper[4578]: I1003 12:52:35.910889 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:35Z","lastTransitionTime":"2025-10-03T12:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.013608 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.013683 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.013694 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.013709 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.013718 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:36Z","lastTransitionTime":"2025-10-03T12:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.116102 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.116135 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.116145 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.116160 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.116169 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:36Z","lastTransitionTime":"2025-10-03T12:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.219454 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.219519 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.219531 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.219576 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.219589 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:36Z","lastTransitionTime":"2025-10-03T12:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.322487 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.322525 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.322534 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.322547 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.322558 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:36Z","lastTransitionTime":"2025-10-03T12:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.424478 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.424511 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.424523 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.424539 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.424549 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:36Z","lastTransitionTime":"2025-10-03T12:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.526982 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.527017 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.527027 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.527041 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.527052 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:36Z","lastTransitionTime":"2025-10-03T12:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.629653 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.629684 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.629699 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.629714 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.629724 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:36Z","lastTransitionTime":"2025-10-03T12:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.732176 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.732214 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.732225 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.732243 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.732260 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:36Z","lastTransitionTime":"2025-10-03T12:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.833933 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.833993 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.834003 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.834020 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.834032 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:36Z","lastTransitionTime":"2025-10-03T12:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.908895 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.908941 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:36 crc kubenswrapper[4578]: E1003 12:52:36.909063 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:36 crc kubenswrapper[4578]: E1003 12:52:36.909139 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.935820 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.936055 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.936119 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.936190 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:36 crc kubenswrapper[4578]: I1003 12:52:36.936248 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:36Z","lastTransitionTime":"2025-10-03T12:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.038061 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.038097 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.038106 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.038121 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.038157 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:37Z","lastTransitionTime":"2025-10-03T12:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.139862 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.139907 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.139918 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.139933 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.139945 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:37Z","lastTransitionTime":"2025-10-03T12:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.241785 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.242096 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.242182 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.242254 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.242313 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:37Z","lastTransitionTime":"2025-10-03T12:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.344666 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.344710 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.344722 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.344738 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.344749 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:37Z","lastTransitionTime":"2025-10-03T12:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.446433 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.446465 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.446475 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.446491 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.446501 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:37Z","lastTransitionTime":"2025-10-03T12:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.548728 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.548766 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.548777 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.548793 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.548874 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:37Z","lastTransitionTime":"2025-10-03T12:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.650906 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.650947 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.650957 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.650972 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.650982 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:37Z","lastTransitionTime":"2025-10-03T12:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.753572 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.753600 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.753652 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.753665 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.753725 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:37Z","lastTransitionTime":"2025-10-03T12:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.855773 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.855808 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.855819 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.855833 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.855846 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:37Z","lastTransitionTime":"2025-10-03T12:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.908988 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.909044 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:37 crc kubenswrapper[4578]: E1003 12:52:37.909128 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:37 crc kubenswrapper[4578]: E1003 12:52:37.909185 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.957504 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.957551 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.957566 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.957583 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:37 crc kubenswrapper[4578]: I1003 12:52:37.957594 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:37Z","lastTransitionTime":"2025-10-03T12:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.059684 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.059721 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.059729 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.059745 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.059755 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:38Z","lastTransitionTime":"2025-10-03T12:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.162504 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.162547 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.162557 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.162571 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.162584 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:38Z","lastTransitionTime":"2025-10-03T12:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.265098 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.265147 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.265158 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.265175 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.265185 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:38Z","lastTransitionTime":"2025-10-03T12:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.367315 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.367351 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.367360 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.367375 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.367385 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:38Z","lastTransitionTime":"2025-10-03T12:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.395766 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.395818 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.395830 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.395845 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.395858 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:38Z","lastTransitionTime":"2025-10-03T12:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:38 crc kubenswrapper[4578]: E1003 12:52:38.409774 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:38Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.413258 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.413290 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.413297 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.413311 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.413319 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:38Z","lastTransitionTime":"2025-10-03T12:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:38 crc kubenswrapper[4578]: E1003 12:52:38.424173 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:38Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.427032 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.427054 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.427062 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.427074 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.427083 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:38Z","lastTransitionTime":"2025-10-03T12:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:38 crc kubenswrapper[4578]: E1003 12:52:38.438113 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:38Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.442180 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.442207 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.442215 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.442228 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.442238 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:38Z","lastTransitionTime":"2025-10-03T12:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:38 crc kubenswrapper[4578]: E1003 12:52:38.455788 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:38Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.459318 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.459369 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.459378 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.459392 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.459401 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:38Z","lastTransitionTime":"2025-10-03T12:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:38 crc kubenswrapper[4578]: E1003 12:52:38.471957 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:38Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:38 crc kubenswrapper[4578]: E1003 12:52:38.472123 4578 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.473816 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.473854 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.473873 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.473895 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.473908 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:38Z","lastTransitionTime":"2025-10-03T12:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.576119 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.576151 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.576159 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.576172 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.576182 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:38Z","lastTransitionTime":"2025-10-03T12:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.678275 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.678322 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.678330 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.678346 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.678357 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:38Z","lastTransitionTime":"2025-10-03T12:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.781186 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.781221 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.781230 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.781245 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.781258 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:38Z","lastTransitionTime":"2025-10-03T12:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.883312 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.883364 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.883374 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.883387 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.883397 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:38Z","lastTransitionTime":"2025-10-03T12:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.909033 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.909077 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:38 crc kubenswrapper[4578]: E1003 12:52:38.909162 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:38 crc kubenswrapper[4578]: E1003 12:52:38.909317 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.985803 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.986092 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.986214 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.986323 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:38 crc kubenswrapper[4578]: I1003 12:52:38.986456 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:38Z","lastTransitionTime":"2025-10-03T12:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.088494 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.088620 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.088669 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.088686 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.088694 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:39Z","lastTransitionTime":"2025-10-03T12:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.191346 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.191375 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.191383 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.191395 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.191403 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:39Z","lastTransitionTime":"2025-10-03T12:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.293229 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.293267 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.293280 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.293301 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.293596 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:39Z","lastTransitionTime":"2025-10-03T12:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.395314 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.395348 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.395356 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.395370 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.395379 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:39Z","lastTransitionTime":"2025-10-03T12:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.497450 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.497484 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.497494 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.497509 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.497520 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:39Z","lastTransitionTime":"2025-10-03T12:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.599678 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.599744 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.599754 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.599767 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.599777 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:39Z","lastTransitionTime":"2025-10-03T12:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.702106 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.702147 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.702155 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.702168 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.702177 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:39Z","lastTransitionTime":"2025-10-03T12:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.805091 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.805132 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.805143 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.805157 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.805167 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:39Z","lastTransitionTime":"2025-10-03T12:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.907582 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.907621 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.907646 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.907663 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.907675 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:39Z","lastTransitionTime":"2025-10-03T12:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.908174 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:39 crc kubenswrapper[4578]: I1003 12:52:39.908187 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:39 crc kubenswrapper[4578]: E1003 12:52:39.908504 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:39 crc kubenswrapper[4578]: E1003 12:52:39.908665 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.010270 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.010302 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.010310 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.010325 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.010334 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:40Z","lastTransitionTime":"2025-10-03T12:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.112961 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.113049 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.113069 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.113092 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.113110 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:40Z","lastTransitionTime":"2025-10-03T12:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.216740 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.217088 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.217105 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.217508 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.217531 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:40Z","lastTransitionTime":"2025-10-03T12:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.324738 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.324987 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.325079 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.325473 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.325575 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:40Z","lastTransitionTime":"2025-10-03T12:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.428125 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.428390 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.428486 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.428605 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.428691 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:40Z","lastTransitionTime":"2025-10-03T12:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.531132 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.531182 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.531191 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.531204 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.531212 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:40Z","lastTransitionTime":"2025-10-03T12:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.632886 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.632928 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.632937 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.632949 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.632957 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:40Z","lastTransitionTime":"2025-10-03T12:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.735298 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.735885 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.735955 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.736034 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.736118 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:40Z","lastTransitionTime":"2025-10-03T12:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.838670 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.838711 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.838723 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.838739 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.838750 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:40Z","lastTransitionTime":"2025-10-03T12:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.908548 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.908592 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:40 crc kubenswrapper[4578]: E1003 12:52:40.909482 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:40 crc kubenswrapper[4578]: E1003 12:52:40.909658 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.941591 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.941668 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.941677 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.941693 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:40 crc kubenswrapper[4578]: I1003 12:52:40.941702 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:40Z","lastTransitionTime":"2025-10-03T12:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.045331 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.045374 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.045385 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.045402 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.045412 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:41Z","lastTransitionTime":"2025-10-03T12:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.148359 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.148402 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.148412 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.148430 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.148440 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:41Z","lastTransitionTime":"2025-10-03T12:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.251681 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.251734 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.251756 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.251786 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.251809 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:41Z","lastTransitionTime":"2025-10-03T12:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.354540 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.354660 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.354674 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.354689 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.354700 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:41Z","lastTransitionTime":"2025-10-03T12:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.457298 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.457334 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.457343 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.457358 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.457368 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:41Z","lastTransitionTime":"2025-10-03T12:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.559141 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.559187 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.559196 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.559222 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.559231 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:41Z","lastTransitionTime":"2025-10-03T12:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.661817 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.661853 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.661864 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.661879 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.661888 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:41Z","lastTransitionTime":"2025-10-03T12:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.765207 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.765257 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.765270 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.765288 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.765300 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:41Z","lastTransitionTime":"2025-10-03T12:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.867347 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.867385 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.867396 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.867410 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.867420 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:41Z","lastTransitionTime":"2025-10-03T12:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.908959 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.908968 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:41 crc kubenswrapper[4578]: E1003 12:52:41.909125 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:41 crc kubenswrapper[4578]: E1003 12:52:41.909186 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.969363 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.969403 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.969413 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.969426 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:41 crc kubenswrapper[4578]: I1003 12:52:41.969434 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:41Z","lastTransitionTime":"2025-10-03T12:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.075422 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.075503 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.075748 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.075769 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.075781 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:42Z","lastTransitionTime":"2025-10-03T12:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.179314 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.179367 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.179378 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.179397 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.179409 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:42Z","lastTransitionTime":"2025-10-03T12:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.281817 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.281857 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.281865 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.281881 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.281891 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:42Z","lastTransitionTime":"2025-10-03T12:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.385713 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.385788 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.385798 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.385812 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.385822 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:42Z","lastTransitionTime":"2025-10-03T12:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.488103 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.488151 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.488163 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.488180 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.488191 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:42Z","lastTransitionTime":"2025-10-03T12:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.590131 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.590171 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.590181 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.590196 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.590206 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:42Z","lastTransitionTime":"2025-10-03T12:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.691899 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.691932 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.691942 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.691956 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.691966 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:42Z","lastTransitionTime":"2025-10-03T12:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.794383 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.794430 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.794444 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.794462 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.794475 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:42Z","lastTransitionTime":"2025-10-03T12:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.896548 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.896583 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.896593 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.896606 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.896616 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:42Z","lastTransitionTime":"2025-10-03T12:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.908610 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:42 crc kubenswrapper[4578]: E1003 12:52:42.908742 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.908900 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:42 crc kubenswrapper[4578]: E1003 12:52:42.909193 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.998546 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.998582 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.998592 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.998607 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:42 crc kubenswrapper[4578]: I1003 12:52:42.998617 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:42Z","lastTransitionTime":"2025-10-03T12:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.100962 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.101001 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.101015 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.101029 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.101037 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:43Z","lastTransitionTime":"2025-10-03T12:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.202869 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.202933 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.202942 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.202956 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.202966 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:43Z","lastTransitionTime":"2025-10-03T12:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.305493 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.305536 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.305554 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.305569 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.305579 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:43Z","lastTransitionTime":"2025-10-03T12:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.407762 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.407821 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.407831 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.407846 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.407856 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:43Z","lastTransitionTime":"2025-10-03T12:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.509901 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.509935 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.509946 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.509961 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.509973 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:43Z","lastTransitionTime":"2025-10-03T12:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.612712 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.612748 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.612760 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.612775 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.612785 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:43Z","lastTransitionTime":"2025-10-03T12:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.715585 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.715648 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.715659 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.715698 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.715709 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:43Z","lastTransitionTime":"2025-10-03T12:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.817477 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.817514 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.817523 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.817536 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.817547 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:43Z","lastTransitionTime":"2025-10-03T12:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.908549 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.908820 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:43 crc kubenswrapper[4578]: E1003 12:52:43.908974 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:43 crc kubenswrapper[4578]: E1003 12:52:43.909022 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.920233 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.920267 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.920276 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.920293 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.920303 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:43Z","lastTransitionTime":"2025-10-03T12:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:43 crc kubenswrapper[4578]: I1003 12:52:43.921715 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.022543 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.022585 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.022595 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.022613 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.022623 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:44Z","lastTransitionTime":"2025-10-03T12:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.126014 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.126069 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.126098 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.126119 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.126129 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:44Z","lastTransitionTime":"2025-10-03T12:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.228487 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.228527 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.228538 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.228585 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.228594 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:44Z","lastTransitionTime":"2025-10-03T12:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.330826 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.330865 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.330877 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.330892 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.330904 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:44Z","lastTransitionTime":"2025-10-03T12:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.432858 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.432920 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.432935 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.432952 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.433308 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:44Z","lastTransitionTime":"2025-10-03T12:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.535984 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.536012 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.536020 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.536032 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.536041 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:44Z","lastTransitionTime":"2025-10-03T12:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.638359 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.638388 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.638396 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.638408 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.638419 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:44Z","lastTransitionTime":"2025-10-03T12:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.741038 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.741072 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.741084 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.741099 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.741111 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:44Z","lastTransitionTime":"2025-10-03T12:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.844015 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.844063 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.844073 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.844087 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.844098 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:44Z","lastTransitionTime":"2025-10-03T12:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.908256 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.908348 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:44 crc kubenswrapper[4578]: E1003 12:52:44.908763 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.909082 4578 scope.go:117] "RemoveContainer" containerID="508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c" Oct 03 12:52:44 crc kubenswrapper[4578]: E1003 12:52:44.909270 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" Oct 03 12:52:44 crc kubenswrapper[4578]: E1003 12:52:44.909329 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.920561 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.926758 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b4650253599f9031287cc43642dff608baa3920e6dce60b433a7c642cece88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.946540 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f593b81-12c2-4633-b4d2-651fac765c95\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://780ac093760f0fec76eae4476b186e59dae84112d66c908fd1a023c235fcd8e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b89889e71c0f28841554cbe4259d4c6e176a5e2cfd7704c046cabde97620444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7fed25a1a59a4c01360e4a4d5a48902dadba7455d77cbcc5905916972f8898e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce12d0f8a591541a8bdf8c68c9dd153e21e01dfd6729a99e2a7227ae12a24285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49927d3d11de7a2eff961d5cc4a6805a23dee422fa83f2c5ddd884ca8d2de0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d4204206079bd1a7dbc4c6dea30c0cbdfbea2e19d2e3cdaa3832bb3d6adacfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4204206079bd1a7dbc4c6dea30c0cbdfbea2e19d2e3cdaa3832bb3d6adacfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69c39d03a71e9b2566a1b59d75b22956ddd2c505fcfe381cae908d50a0590ef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69c39d03a71e9b2566a1b59d75b22956ddd2c505fcfe381cae908d50a0590ef8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47366a9654db257f55811ac4484b3f86532df669c562b68f597b938f322d916f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47366a9654db257f55811ac4484b3f86532df669c562b68f597b938f322d916f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.949514 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.949850 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.949952 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.950067 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.950167 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:44Z","lastTransitionTime":"2025-10-03T12:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.970334 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:44 crc kubenswrapper[4578]: I1003 12:52:44.988484 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:44Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.023621 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:29Z\\\",\\\"message\\\":\\\"0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:52:29.037364 6557 services_controller.go:444] Built service openshift-machine-config-operator/machine-config-controller LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1003 12:52:29.037373 6557 services_controller.go:445] Built service openshift-machine-config-operator/machine-config-controller LB template configs for network=default: []services.lbConfig(nil)\\\\nI1003 12:52:29.037391 6557 services_controller.go:451] Built service openshift-machine-config-operator/machine-config-controller cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.16\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Router\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:52:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.042541 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.052440 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.052766 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.052857 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.052937 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.053003 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:45Z","lastTransitionTime":"2025-10-03T12:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.054217 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e43f233-de60-499f-bcc6-8e782a42f31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ede2cac008ac6d78a301b36ea4cdd25ccc11646c529d397c230d264590fcbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60be5924c0b93bbd78ef07fc29492fd17b85c06407edcecfb7763032326000b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60be5924c0b93bbd78ef07fc29492fd17b85c06407edcecfb7763032326000b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.068022 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650520feb6cfcf5c6aea04b4a990593eab6f2a2ddec00e2270105f25ee9a067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:21Z\\\",\\\"message\\\":\\\"2025-10-03T12:51:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_57184fea-0a56-4895-9b4d-dbeabc7e5d55\\\\n2025-10-03T12:51:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_57184fea-0a56-4895-9b4d-dbeabc7e5d55 to /host/opt/cni/bin/\\\\n2025-10-03T12:51:36Z [verbose] multus-daemon started\\\\n2025-10-03T12:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-03T12:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.080424 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.090972 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.102420 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a524da53-3b46-423d-97ae-0e7d131b958f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d4f0ede4a9f9262ffe08a7abc4366faf1001b25f3a6a35dfac42c3d8362030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ba8c08b2dceaf712d8d6b27f5c3a4b6fd3a204fc494dfb8ec33ed7ce8bc16e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1fb8c53aa23e2c70af59f270c04d09599d79718098c17a809ace5f72797303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.117259 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.129955 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.142068 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.152931 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.155503 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.155671 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.155757 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.155839 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.155912 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:45Z","lastTransitionTime":"2025-10-03T12:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.166146 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.178925 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.192135 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:45Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.258045 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.258814 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.258849 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.258867 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.258876 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:45Z","lastTransitionTime":"2025-10-03T12:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.360748 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.360781 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.360792 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.360806 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.360816 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:45Z","lastTransitionTime":"2025-10-03T12:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.462793 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.462827 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.462838 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.462853 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.462865 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:45Z","lastTransitionTime":"2025-10-03T12:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.565623 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.565720 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.565734 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.565749 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.565759 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:45Z","lastTransitionTime":"2025-10-03T12:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.668158 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.668192 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.668204 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.668231 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.668244 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:45Z","lastTransitionTime":"2025-10-03T12:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.771616 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.771723 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.771740 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.771772 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.771789 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:45Z","lastTransitionTime":"2025-10-03T12:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.874330 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.874407 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.874421 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.874446 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.874460 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:45Z","lastTransitionTime":"2025-10-03T12:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.908846 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.908883 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:45 crc kubenswrapper[4578]: E1003 12:52:45.908993 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:45 crc kubenswrapper[4578]: E1003 12:52:45.909053 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.977095 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.977164 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.977173 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.977187 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:45 crc kubenswrapper[4578]: I1003 12:52:45.977195 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:45Z","lastTransitionTime":"2025-10-03T12:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.083657 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.083709 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.083722 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.083739 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.083750 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:46Z","lastTransitionTime":"2025-10-03T12:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.186196 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.186237 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.186248 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.186262 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.186273 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:46Z","lastTransitionTime":"2025-10-03T12:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.288745 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.288807 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.288816 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.288828 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.288854 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:46Z","lastTransitionTime":"2025-10-03T12:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.391844 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.392091 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.392212 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.392314 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.392394 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:46Z","lastTransitionTime":"2025-10-03T12:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.496087 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.496162 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.496174 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.496193 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.496205 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:46Z","lastTransitionTime":"2025-10-03T12:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.598779 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.599070 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.599215 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.599289 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.599379 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:46Z","lastTransitionTime":"2025-10-03T12:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.702259 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.702316 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.702326 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.702339 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.702347 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:46Z","lastTransitionTime":"2025-10-03T12:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.805214 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.805274 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.805287 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.805307 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.805325 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:46Z","lastTransitionTime":"2025-10-03T12:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.908067 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:46 crc kubenswrapper[4578]: E1003 12:52:46.908210 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.908323 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.908408 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.908418 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.908476 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.908498 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:46Z","lastTransitionTime":"2025-10-03T12:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:46 crc kubenswrapper[4578]: I1003 12:52:46.909396 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:46 crc kubenswrapper[4578]: E1003 12:52:46.909541 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.011439 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.011484 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.011495 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.011512 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.011524 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:47Z","lastTransitionTime":"2025-10-03T12:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.114430 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.114472 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.114482 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.114534 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.114548 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:47Z","lastTransitionTime":"2025-10-03T12:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.217753 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.217831 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.217863 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.217896 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.218015 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:47Z","lastTransitionTime":"2025-10-03T12:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.322690 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.322775 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.322794 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.322824 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.322848 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:47Z","lastTransitionTime":"2025-10-03T12:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.427328 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.427374 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.427387 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.427421 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.427437 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:47Z","lastTransitionTime":"2025-10-03T12:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.530126 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.530167 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.530177 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.530193 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.530206 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:47Z","lastTransitionTime":"2025-10-03T12:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.632790 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.632859 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.632868 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.632883 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.632893 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:47Z","lastTransitionTime":"2025-10-03T12:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.734909 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.735177 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.735274 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.735369 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.735434 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:47Z","lastTransitionTime":"2025-10-03T12:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.837815 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.838204 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.838327 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.838499 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.838614 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:47Z","lastTransitionTime":"2025-10-03T12:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.908418 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:47 crc kubenswrapper[4578]: E1003 12:52:47.908594 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.908428 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:47 crc kubenswrapper[4578]: E1003 12:52:47.908732 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.941527 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.941560 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.941569 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.941591 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:47 crc kubenswrapper[4578]: I1003 12:52:47.941601 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:47Z","lastTransitionTime":"2025-10-03T12:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.044271 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.044352 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.044374 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.044403 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.044464 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:48Z","lastTransitionTime":"2025-10-03T12:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.148115 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.148195 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.148214 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.148264 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.148279 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:48Z","lastTransitionTime":"2025-10-03T12:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.251588 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.251722 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.251739 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.251771 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.251788 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:48Z","lastTransitionTime":"2025-10-03T12:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.353796 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.353834 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.353844 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.353858 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.353870 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:48Z","lastTransitionTime":"2025-10-03T12:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.456386 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.456428 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.456438 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.456453 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.456463 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:48Z","lastTransitionTime":"2025-10-03T12:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.528586 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.528860 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.528949 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.529025 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.529086 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:48Z","lastTransitionTime":"2025-10-03T12:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:48 crc kubenswrapper[4578]: E1003 12:52:48.541224 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:48Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.547872 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.547925 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.547936 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.547949 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.547958 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:48Z","lastTransitionTime":"2025-10-03T12:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:48 crc kubenswrapper[4578]: E1003 12:52:48.559168 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:48Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.562076 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.562238 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.562319 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.562381 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.562438 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:48Z","lastTransitionTime":"2025-10-03T12:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:48 crc kubenswrapper[4578]: E1003 12:52:48.573738 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:48Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.576397 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.576541 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.576664 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.576753 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.576847 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:48Z","lastTransitionTime":"2025-10-03T12:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:48 crc kubenswrapper[4578]: E1003 12:52:48.587820 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:48Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.590904 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.590923 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.590931 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.590943 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.590951 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:48Z","lastTransitionTime":"2025-10-03T12:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:48 crc kubenswrapper[4578]: E1003 12:52:48.602149 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:48Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:48 crc kubenswrapper[4578]: E1003 12:52:48.602585 4578 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.604176 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.604194 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.604201 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.604212 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.604222 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:48Z","lastTransitionTime":"2025-10-03T12:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.707038 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.707077 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.707087 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.707107 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.707125 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:48Z","lastTransitionTime":"2025-10-03T12:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.809134 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.809451 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.809554 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.809657 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.809728 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:48Z","lastTransitionTime":"2025-10-03T12:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.908895 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:48 crc kubenswrapper[4578]: E1003 12:52:48.909569 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.908895 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:48 crc kubenswrapper[4578]: E1003 12:52:48.910008 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.912481 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.912508 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.912518 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.912532 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:48 crc kubenswrapper[4578]: I1003 12:52:48.912543 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:48Z","lastTransitionTime":"2025-10-03T12:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.014727 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.014774 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.014786 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.014801 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.014812 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:49Z","lastTransitionTime":"2025-10-03T12:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.117658 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.117744 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.117773 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.117803 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.117825 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:49Z","lastTransitionTime":"2025-10-03T12:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.220800 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.220842 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.220854 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.220869 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.220880 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:49Z","lastTransitionTime":"2025-10-03T12:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.280465 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs\") pod \"network-metrics-daemon-q8986\" (UID: \"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\") " pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:49 crc kubenswrapper[4578]: E1003 12:52:49.280745 4578 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 12:52:49 crc kubenswrapper[4578]: E1003 12:52:49.280807 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs podName:2971c7bb-e0f4-44a8-8f52-9ea8ba199772 nodeName:}" failed. No retries permitted until 2025-10-03 12:53:53.280789383 +0000 UTC m=+169.079261567 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs") pod "network-metrics-daemon-q8986" (UID: "2971c7bb-e0f4-44a8-8f52-9ea8ba199772") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.323472 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.323513 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.323522 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.323537 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.323547 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:49Z","lastTransitionTime":"2025-10-03T12:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.426514 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.426556 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.426568 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.426584 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.426596 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:49Z","lastTransitionTime":"2025-10-03T12:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.530106 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.530317 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.530404 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.530468 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.530522 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:49Z","lastTransitionTime":"2025-10-03T12:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.633325 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.633371 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.633379 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.633393 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.633403 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:49Z","lastTransitionTime":"2025-10-03T12:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.735787 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.735815 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.735824 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.735836 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.735846 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:49Z","lastTransitionTime":"2025-10-03T12:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.837729 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.837765 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.837775 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.837790 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.837801 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:49Z","lastTransitionTime":"2025-10-03T12:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.908433 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:49 crc kubenswrapper[4578]: E1003 12:52:49.908552 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.908716 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:49 crc kubenswrapper[4578]: E1003 12:52:49.908792 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.940255 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.940297 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.940310 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.940324 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:49 crc kubenswrapper[4578]: I1003 12:52:49.940335 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:49Z","lastTransitionTime":"2025-10-03T12:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.042250 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.042531 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.042618 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.042753 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.042841 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:50Z","lastTransitionTime":"2025-10-03T12:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.145659 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.145705 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.145717 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.145730 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.145739 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:50Z","lastTransitionTime":"2025-10-03T12:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.247547 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.247828 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.247893 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.247970 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.248031 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:50Z","lastTransitionTime":"2025-10-03T12:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.350251 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.350485 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.350579 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.350665 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.350738 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:50Z","lastTransitionTime":"2025-10-03T12:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.452805 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.452843 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.452851 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.452865 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.452874 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:50Z","lastTransitionTime":"2025-10-03T12:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.555118 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.555374 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.555519 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.555699 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.555889 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:50Z","lastTransitionTime":"2025-10-03T12:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.658341 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.658571 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.658677 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.658754 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.658897 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:50Z","lastTransitionTime":"2025-10-03T12:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.761112 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.761394 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.761519 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.761708 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.761909 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:50Z","lastTransitionTime":"2025-10-03T12:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.865096 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.865137 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.865148 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.865162 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.865170 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:50Z","lastTransitionTime":"2025-10-03T12:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.908328 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.908328 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:50 crc kubenswrapper[4578]: E1003 12:52:50.908806 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:50 crc kubenswrapper[4578]: E1003 12:52:50.908883 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.967620 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.968088 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.968206 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.968302 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:50 crc kubenswrapper[4578]: I1003 12:52:50.968379 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:50Z","lastTransitionTime":"2025-10-03T12:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.070280 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.070338 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.070351 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.070366 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.070376 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:51Z","lastTransitionTime":"2025-10-03T12:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.172697 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.172993 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.173064 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.173148 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.173220 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:51Z","lastTransitionTime":"2025-10-03T12:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.275616 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.275881 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.275998 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.276113 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.276211 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:51Z","lastTransitionTime":"2025-10-03T12:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.378949 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.378995 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.379006 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.379021 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.379034 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:51Z","lastTransitionTime":"2025-10-03T12:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.481406 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.481453 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.481465 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.481482 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.481494 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:51Z","lastTransitionTime":"2025-10-03T12:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.583953 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.584004 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.584013 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.584027 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.584037 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:51Z","lastTransitionTime":"2025-10-03T12:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.685940 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.686163 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.686254 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.686326 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.686388 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:51Z","lastTransitionTime":"2025-10-03T12:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.788402 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.788457 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.788469 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.788483 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.788493 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:51Z","lastTransitionTime":"2025-10-03T12:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.890849 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.891089 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.891191 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.891286 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.891350 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:51Z","lastTransitionTime":"2025-10-03T12:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.908427 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.908427 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:51 crc kubenswrapper[4578]: E1003 12:52:51.908798 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:51 crc kubenswrapper[4578]: E1003 12:52:51.908813 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.993183 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.993213 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.993222 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.993233 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:51 crc kubenswrapper[4578]: I1003 12:52:51.993241 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:51Z","lastTransitionTime":"2025-10-03T12:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.095419 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.095457 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.095466 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.095480 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.095489 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:52Z","lastTransitionTime":"2025-10-03T12:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.198673 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.198725 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.198737 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.198751 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.198764 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:52Z","lastTransitionTime":"2025-10-03T12:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.301733 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.301773 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.301783 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.301798 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.301809 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:52Z","lastTransitionTime":"2025-10-03T12:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.404132 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.404159 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.404166 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.404179 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.404187 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:52Z","lastTransitionTime":"2025-10-03T12:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.506858 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.507117 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.507249 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.507347 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.507453 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:52Z","lastTransitionTime":"2025-10-03T12:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.610947 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.611268 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.611378 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.611471 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.611552 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:52Z","lastTransitionTime":"2025-10-03T12:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.715452 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.715887 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.716017 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.716088 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.716188 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:52Z","lastTransitionTime":"2025-10-03T12:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.819060 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.819093 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.819101 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.819114 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.819124 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:52Z","lastTransitionTime":"2025-10-03T12:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.908769 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:52 crc kubenswrapper[4578]: E1003 12:52:52.908878 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.910295 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:52 crc kubenswrapper[4578]: E1003 12:52:52.910403 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.920482 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.920538 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.920550 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.920570 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:52 crc kubenswrapper[4578]: I1003 12:52:52.920581 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:52Z","lastTransitionTime":"2025-10-03T12:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.023489 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.024160 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.024252 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.024346 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.024434 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:53Z","lastTransitionTime":"2025-10-03T12:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.127877 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.128016 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.128027 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.128045 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.128058 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:53Z","lastTransitionTime":"2025-10-03T12:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.230999 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.231031 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.231039 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.231052 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.231061 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:53Z","lastTransitionTime":"2025-10-03T12:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.333005 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.333080 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.333091 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.333106 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.333122 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:53Z","lastTransitionTime":"2025-10-03T12:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.436035 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.436086 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.436096 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.436110 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.436120 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:53Z","lastTransitionTime":"2025-10-03T12:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.538992 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.539025 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.539033 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.539045 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.539055 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:53Z","lastTransitionTime":"2025-10-03T12:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.641180 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.641246 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.641260 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.641276 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.641287 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:53Z","lastTransitionTime":"2025-10-03T12:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.743906 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.743973 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.743986 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.744001 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.744032 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:53Z","lastTransitionTime":"2025-10-03T12:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.846169 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.846221 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.846234 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.846252 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.846265 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:53Z","lastTransitionTime":"2025-10-03T12:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.908685 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:53 crc kubenswrapper[4578]: E1003 12:52:53.908819 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.908871 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:53 crc kubenswrapper[4578]: E1003 12:52:53.908988 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.948989 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.949021 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.949029 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.949042 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:53 crc kubenswrapper[4578]: I1003 12:52:53.949051 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:53Z","lastTransitionTime":"2025-10-03T12:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.051541 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.051582 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.051591 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.051606 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.051616 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:54Z","lastTransitionTime":"2025-10-03T12:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.154089 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.154322 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.154405 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.154512 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.154586 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:54Z","lastTransitionTime":"2025-10-03T12:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.257015 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.257049 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.257058 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.257072 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.257094 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:54Z","lastTransitionTime":"2025-10-03T12:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.359800 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.359840 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.359850 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.359865 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.359876 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:54Z","lastTransitionTime":"2025-10-03T12:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.461804 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.461862 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.461880 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.461905 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.461924 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:54Z","lastTransitionTime":"2025-10-03T12:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.564409 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.564455 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.564468 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.564485 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.564500 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:54Z","lastTransitionTime":"2025-10-03T12:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.666934 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.666969 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.666980 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.666997 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.667006 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:54Z","lastTransitionTime":"2025-10-03T12:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.769416 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.769676 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.769769 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.769920 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.770018 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:54Z","lastTransitionTime":"2025-10-03T12:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.872395 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.872435 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.872448 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.872463 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.872475 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:54Z","lastTransitionTime":"2025-10-03T12:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.908797 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:54 crc kubenswrapper[4578]: E1003 12:52:54.908954 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.908805 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:54 crc kubenswrapper[4578]: E1003 12:52:54.909438 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.922359 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d64423fc-1a76-48fd-a3c3-74a64e791db5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c39656adf29002ba3e5c419f80a8cf68c8f7ddf9245ba004d0aa93930fbabece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb906d24f8d6787738a916592c2d2aed500fa5b1b5f7366b2f836e9a69d18e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d63272119a51846872a26ee010f0edb31fbafdb189de6425227d750de1634d76\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40f9f5d7ef3a969125784cd3618c25152b78d9f8f554241f1b69d24f317ee2ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.932811 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e43f233-de60-499f-bcc6-8e782a42f31f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5ede2cac008ac6d78a301b36ea4cdd25ccc11646c529d397c230d264590fcbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60be5924c0b93bbd78ef07fc29492fd17b85c06407edcecfb7763032326000b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60be5924c0b93bbd78ef07fc29492fd17b85c06407edcecfb7763032326000b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.946581 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-c4jgn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fed3968f-4bb0-4209-a377-16c26f9ea070\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://650520feb6cfcf5c6aea04b4a990593eab6f2a2ddec00e2270105f25ee9a067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:21Z\\\",\\\"message\\\":\\\"2025-10-03T12:51:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_57184fea-0a56-4895-9b4d-dbeabc7e5d55\\\\n2025-10-03T12:51:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_57184fea-0a56-4895-9b4d-dbeabc7e5d55 to /host/opt/cni/bin/\\\\n2025-10-03T12:51:36Z [verbose] multus-daemon started\\\\n2025-10-03T12:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-03T12:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:52:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qxn6b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-c4jgn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.957814 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b7d7e663-26ba-48e2-849e-f7c9554fd7a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5d355922e85cf9fb2f42a76d46c58d4da6ba4ef8d3957feb5f0a2d64c40d053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nkhmc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-qq9jm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.967394 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-rh9f2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12c73964-5ff0-4b3a-8ec6-774035e92340\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e667f32ce7426f6ccb9e395a0a6b0b6b65e2d446bba60c9b1b30c447d31e125d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zf7m6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:34Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-rh9f2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.974735 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.974761 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.974768 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.974782 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.974791 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:54Z","lastTransitionTime":"2025-10-03T12:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.978586 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a524da53-3b46-423d-97ae-0e7d131b958f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6d4f0ede4a9f9262ffe08a7abc4366faf1001b25f3a6a35dfac42c3d8362030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ba8c08b2dceaf712d8d6b27f5c3a4b6fd3a204fc494dfb8ec33ed7ce8bc16e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d1fb8c53aa23e2c70af59f270c04d09599d79718098c17a809ace5f72797303\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faae1272af41b608e1f73adf5bc9645378cb8a2318b108d1e47908775b249afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:54 crc kubenswrapper[4578]: I1003 12:52:54.989095 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46b16f6ad654949a86f494ebeea845cdb5289ca9327f6d6f79e4894724dc27ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4520f9aa89e80a3abe9cd2f9a2f3bdb24e158e73de575b2a36c30a490c6aae12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.000250 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:54Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.008206 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p8d2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbf63262-15f9-4fe4-b6f7-a3fc69bc07ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25df054cb00bd9e8d4ed457486973e29d6f6b018ed4ec45427701d3bbb688d50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qvxhl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p8d2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:55Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.021692 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e95e1c26-64e1-467b-bf8d-aa066847a95a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://385d8d5723709083588d67cb8b13d292874f0775cc51b44fd44b2ce1f63d75a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38bd0957bc046426bdf77235a032a0221ebf22010f61ed7fd5c717cfe8d22177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sr4tx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xc4j9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:55Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.031427 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0637465c862f96d73aa49a9342b8c7dc1612d7a63da02de133ec91e4bed6d46d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:55Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.040299 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:55Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.051707 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5403769c-f73e-4921-a402-3d650774c2e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://586408613c47042daefa90869c5a4dd44f6e32b931a478c5496e8d2d13b145c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://838746e7370d0aee3911c78845bd9a131ec12af47aedd80b3fd34120a0081f29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0df32050f39aba0ed66c6beaa8f864c4b8e7746c9734a5252344100818a03da9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e03e8c17e80f93f27646c1e0ba12cacd8e42c37aba0612faed308c4bff977bc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://625df44221bc17cbf00dd2355b0eb8b2f5bcb5d195f30791601b19451c92163c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5f7783c18a35bffcb19eb68b8f05ede372d9ffc01ba32c3e5c2ffd1c7c0114e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f40cc4d92a39a774d4a9bc4c07d3ae96b01cec7318bb79d6879d1498d32198b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrtvm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jjvmh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:55Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.061249 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-q8986" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x9d9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-q8986\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:55Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.072602 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abea6eb2-9108-43dd-88c3-c2fedef72f58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1607ec25133cef4404d971b1cd49afbdcecfaafb1c0d449f17aac6784de696af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cce94bf225f930371cdaef023ef1c10d7539b9386dfe642e64ae5a280ff29dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9a351b7b73cf607fb7433039213ceb43a6e4379bb3080cc6464808965049294\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93b4650253599f9031287cc43642dff608baa3920e6dce60b433a7c642cece88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9378a075434f85e592c61410666d214d39b9f932dc5659932f6c63410da90cb1\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-03T12:51:29Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1003 12:51:23.788293 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1003 12:51:23.789037 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3015095834/tls.crt::/tmp/serving-cert-3015095834/tls.key\\\\\\\"\\\\nI1003 12:51:29.545787 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1003 12:51:29.548461 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1003 12:51:29.548481 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1003 12:51:29.548501 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1003 12:51:29.548507 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1003 12:51:29.563074 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1003 12:51:29.563375 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1003 12:51:29.565285 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1003 12:51:29.566357 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1003 12:51:29.566391 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1003 12:51:29.566397 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1003 12:51:29.566409 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1003 12:51:29.566415 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1003 12:51:29.566774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12d1dcc3be921e73a40090992384f6b275a2a416c48aa5196c08dab428dc6daf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8c07a2563c1da400f21b0cc19fa423cc67c3ee58166d8de1adddb571e560d6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:55Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.076588 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.076823 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.077051 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.077254 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.077439 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:55Z","lastTransitionTime":"2025-10-03T12:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.093219 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6f593b81-12c2-4633-b4d2-651fac765c95\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://780ac093760f0fec76eae4476b186e59dae84112d66c908fd1a023c235fcd8e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b89889e71c0f28841554cbe4259d4c6e176a5e2cfd7704c046cabde97620444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7fed25a1a59a4c01360e4a4d5a48902dadba7455d77cbcc5905916972f8898e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce12d0f8a591541a8bdf8c68c9dd153e21e01dfd6729a99e2a7227ae12a24285\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://49927d3d11de7a2eff961d5cc4a6805a23dee422fa83f2c5ddd884ca8d2de0df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d4204206079bd1a7dbc4c6dea30c0cbdfbea2e19d2e3cdaa3832bb3d6adacfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d4204206079bd1a7dbc4c6dea30c0cbdfbea2e19d2e3cdaa3832bb3d6adacfb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69c39d03a71e9b2566a1b59d75b22956ddd2c505fcfe381cae908d50a0590ef8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69c39d03a71e9b2566a1b59d75b22956ddd2c505fcfe381cae908d50a0590ef8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47366a9654db257f55811ac4484b3f86532df669c562b68f597b938f322d916f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47366a9654db257f55811ac4484b3f86532df669c562b68f597b938f322d916f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:55Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.109570 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6347989cc9776189ecd90eccdfbc1dd3dd1358affcec61e397917415a44e6734\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:55Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.121553 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:55Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.141670 4578 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-03T12:52:29Z\\\",\\\"message\\\":\\\"0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1003 12:52:29.037364 6557 services_controller.go:444] Built service openshift-machine-config-operator/machine-config-controller LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1003 12:52:29.037373 6557 services_controller.go:445] Built service openshift-machine-config-operator/machine-config-controller LB template configs for network=default: []services.lbConfig(nil)\\\\nI1003 12:52:29.037391 6557 services_controller.go:451] Built service openshift-machine-config-operator/machine-config-controller cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-controller_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-controller\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.16\\\\\\\", Port:9001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Router\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-03T12:52:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-03T12:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-03T12:51:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-03T12:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gt6c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-03T12:51:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8qvh9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:55Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.180183 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.180215 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.180224 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.180236 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.180244 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:55Z","lastTransitionTime":"2025-10-03T12:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.283119 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.283153 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.283161 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.283174 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.283183 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:55Z","lastTransitionTime":"2025-10-03T12:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.384893 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.384949 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.384959 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.384974 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.384982 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:55Z","lastTransitionTime":"2025-10-03T12:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.487380 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.487459 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.487472 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.487492 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.487504 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:55Z","lastTransitionTime":"2025-10-03T12:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.590044 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.590094 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.590108 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.590124 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.590134 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:55Z","lastTransitionTime":"2025-10-03T12:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.691949 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.692191 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.692270 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.692342 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.692417 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:55Z","lastTransitionTime":"2025-10-03T12:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.794704 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.794741 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.794750 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.794765 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.794775 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:55Z","lastTransitionTime":"2025-10-03T12:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.896672 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.896705 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.896714 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.896727 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.896738 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:55Z","lastTransitionTime":"2025-10-03T12:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.908443 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.908444 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:55 crc kubenswrapper[4578]: E1003 12:52:55.908580 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:55 crc kubenswrapper[4578]: E1003 12:52:55.908997 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.909211 4578 scope.go:117] "RemoveContainer" containerID="508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c" Oct 03 12:52:55 crc kubenswrapper[4578]: E1003 12:52:55.909343 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.998514 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.998556 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.998567 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.998581 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:55 crc kubenswrapper[4578]: I1003 12:52:55.998590 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:55Z","lastTransitionTime":"2025-10-03T12:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.100527 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.100565 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.100575 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.100590 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.100600 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:56Z","lastTransitionTime":"2025-10-03T12:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.202965 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.202998 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.203008 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.203022 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.203031 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:56Z","lastTransitionTime":"2025-10-03T12:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.305410 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.305450 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.305459 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.305473 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.305481 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:56Z","lastTransitionTime":"2025-10-03T12:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.407562 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.407601 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.407613 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.407647 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.407658 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:56Z","lastTransitionTime":"2025-10-03T12:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.509911 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.509956 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.509969 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.509984 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.509994 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:56Z","lastTransitionTime":"2025-10-03T12:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.612546 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.612598 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.612607 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.612620 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.612639 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:56Z","lastTransitionTime":"2025-10-03T12:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.715370 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.715400 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.715409 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.715422 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.715430 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:56Z","lastTransitionTime":"2025-10-03T12:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.817304 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.817353 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.817364 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.817378 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.817387 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:56Z","lastTransitionTime":"2025-10-03T12:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.908839 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.908930 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:56 crc kubenswrapper[4578]: E1003 12:52:56.908996 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:56 crc kubenswrapper[4578]: E1003 12:52:56.909118 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.919468 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.919504 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.919513 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.919527 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:56 crc kubenswrapper[4578]: I1003 12:52:56.919537 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:56Z","lastTransitionTime":"2025-10-03T12:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.021398 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.021443 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.021459 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.021476 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.021488 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:57Z","lastTransitionTime":"2025-10-03T12:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.123362 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.123460 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.123471 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.123487 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.123498 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:57Z","lastTransitionTime":"2025-10-03T12:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.225573 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.225615 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.225649 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.225664 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.225675 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:57Z","lastTransitionTime":"2025-10-03T12:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.327238 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.327277 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.327287 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.327300 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.327309 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:57Z","lastTransitionTime":"2025-10-03T12:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.430170 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.430212 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.430220 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.430234 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.430242 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:57Z","lastTransitionTime":"2025-10-03T12:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.533329 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.533386 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.533394 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.533408 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.533419 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:57Z","lastTransitionTime":"2025-10-03T12:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.636152 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.636191 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.636202 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.636218 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.636229 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:57Z","lastTransitionTime":"2025-10-03T12:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.738911 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.738945 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.738955 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.738967 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.738977 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:57Z","lastTransitionTime":"2025-10-03T12:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.841648 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.841679 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.841690 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.841704 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.841716 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:57Z","lastTransitionTime":"2025-10-03T12:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.908248 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.908296 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:57 crc kubenswrapper[4578]: E1003 12:52:57.908403 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:57 crc kubenswrapper[4578]: E1003 12:52:57.908625 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.944381 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.944420 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.944430 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.944443 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:57 crc kubenswrapper[4578]: I1003 12:52:57.944453 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:57Z","lastTransitionTime":"2025-10-03T12:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.047022 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.047057 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.047067 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.047082 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.047092 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:58Z","lastTransitionTime":"2025-10-03T12:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.149440 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.149472 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.149481 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.149494 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.149503 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:58Z","lastTransitionTime":"2025-10-03T12:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.251961 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.251992 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.252000 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.252012 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.252022 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:58Z","lastTransitionTime":"2025-10-03T12:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.354066 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.354100 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.354110 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.354125 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.354134 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:58Z","lastTransitionTime":"2025-10-03T12:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.456226 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.456260 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.456269 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.456283 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.456291 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:58Z","lastTransitionTime":"2025-10-03T12:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.558486 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.558515 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.558525 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.558540 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.558550 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:58Z","lastTransitionTime":"2025-10-03T12:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.661091 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.661119 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.661127 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.661141 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.661151 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:58Z","lastTransitionTime":"2025-10-03T12:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.763725 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.763811 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.763832 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.763858 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.763875 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:58Z","lastTransitionTime":"2025-10-03T12:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.866366 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.866409 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.866420 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.866436 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.866450 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:58Z","lastTransitionTime":"2025-10-03T12:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.876954 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.877002 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.877015 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.877033 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.877046 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:58Z","lastTransitionTime":"2025-10-03T12:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:58 crc kubenswrapper[4578]: E1003 12:52:58.888570 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:58Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.892045 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.892089 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.892098 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.892111 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.892119 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:58Z","lastTransitionTime":"2025-10-03T12:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:58 crc kubenswrapper[4578]: E1003 12:52:58.905377 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:58Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.908024 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.908024 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:52:58 crc kubenswrapper[4578]: E1003 12:52:58.908175 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:52:58 crc kubenswrapper[4578]: E1003 12:52:58.908279 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.908869 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.908904 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.908914 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.908928 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.908936 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:58Z","lastTransitionTime":"2025-10-03T12:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:58 crc kubenswrapper[4578]: E1003 12:52:58.923031 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:58Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.927047 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.927138 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.927155 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.927175 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.927193 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:58Z","lastTransitionTime":"2025-10-03T12:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:58 crc kubenswrapper[4578]: E1003 12:52:58.941316 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:58Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.945119 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.945169 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.945181 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.945201 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.945215 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:58Z","lastTransitionTime":"2025-10-03T12:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:58 crc kubenswrapper[4578]: E1003 12:52:58.958870 4578 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-03T12:52:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"37bf46c8-7e72-46ed-94e0-9573cda7bef8\\\",\\\"systemUUID\\\":\\\"6a8a1f0a-2ca9-4c9f-abcd-dc137e6c86d8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-03T12:52:58Z is after 2025-08-24T17:21:41Z" Oct 03 12:52:58 crc kubenswrapper[4578]: E1003 12:52:58.959032 4578 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.968441 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.968502 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.968519 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.968538 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:58 crc kubenswrapper[4578]: I1003 12:52:58.968553 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:58Z","lastTransitionTime":"2025-10-03T12:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.071517 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.071558 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.071578 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.071595 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.071614 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:59Z","lastTransitionTime":"2025-10-03T12:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.173696 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.173720 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.173728 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.173740 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.173748 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:59Z","lastTransitionTime":"2025-10-03T12:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.276540 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.276668 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.276682 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.276701 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.276720 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:59Z","lastTransitionTime":"2025-10-03T12:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.379893 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.379962 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.379980 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.380001 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.380016 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:59Z","lastTransitionTime":"2025-10-03T12:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.482743 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.482784 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.482798 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.482817 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.482828 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:59Z","lastTransitionTime":"2025-10-03T12:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.585395 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.585432 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.585441 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.585454 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.585463 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:59Z","lastTransitionTime":"2025-10-03T12:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.687692 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.687754 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.687769 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.687789 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.687804 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:59Z","lastTransitionTime":"2025-10-03T12:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.790172 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.790211 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.790219 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.790232 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.790241 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:59Z","lastTransitionTime":"2025-10-03T12:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.892466 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.892504 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.892512 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.892524 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.892533 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:59Z","lastTransitionTime":"2025-10-03T12:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.908915 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.908943 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:52:59 crc kubenswrapper[4578]: E1003 12:52:59.909130 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:52:59 crc kubenswrapper[4578]: E1003 12:52:59.909233 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.994467 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.994511 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.994523 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.994539 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:52:59 crc kubenswrapper[4578]: I1003 12:52:59.994551 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:52:59Z","lastTransitionTime":"2025-10-03T12:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.097433 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.097500 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.097517 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.097534 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.097549 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:00Z","lastTransitionTime":"2025-10-03T12:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.200126 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.200337 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.200357 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.200382 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.200399 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:00Z","lastTransitionTime":"2025-10-03T12:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.302812 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.302880 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.302893 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.302908 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.302919 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:00Z","lastTransitionTime":"2025-10-03T12:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.405004 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.405141 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.405155 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.405172 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.405184 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:00Z","lastTransitionTime":"2025-10-03T12:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.506935 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.506965 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.506974 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.506986 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.506996 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:00Z","lastTransitionTime":"2025-10-03T12:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.609490 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.609530 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.609539 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.609554 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.609569 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:00Z","lastTransitionTime":"2025-10-03T12:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.712415 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.712471 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.712496 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.712524 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.712546 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:00Z","lastTransitionTime":"2025-10-03T12:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.815669 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.815733 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.815755 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.815785 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.815806 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:00Z","lastTransitionTime":"2025-10-03T12:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.908985 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.909203 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:00 crc kubenswrapper[4578]: E1003 12:53:00.909308 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:53:00 crc kubenswrapper[4578]: E1003 12:53:00.909468 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.917763 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.917799 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.917811 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.917828 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:00 crc kubenswrapper[4578]: I1003 12:53:00.917840 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:00Z","lastTransitionTime":"2025-10-03T12:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.020016 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.020057 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.020070 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.020087 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.020100 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:01Z","lastTransitionTime":"2025-10-03T12:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.122313 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.122353 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.122364 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.122378 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.122386 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:01Z","lastTransitionTime":"2025-10-03T12:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.224877 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.224916 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.224926 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.224941 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.224951 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:01Z","lastTransitionTime":"2025-10-03T12:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.327855 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.327902 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.327914 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.327932 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.327944 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:01Z","lastTransitionTime":"2025-10-03T12:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.429813 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.429841 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.429848 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.429860 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.429872 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:01Z","lastTransitionTime":"2025-10-03T12:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.531373 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.531403 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.531415 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.531429 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.531439 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:01Z","lastTransitionTime":"2025-10-03T12:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.633443 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.633480 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.633491 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.633506 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.633520 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:01Z","lastTransitionTime":"2025-10-03T12:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.736045 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.736099 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.736112 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.736133 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.736148 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:01Z","lastTransitionTime":"2025-10-03T12:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.838817 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.838867 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.838879 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.838897 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.838908 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:01Z","lastTransitionTime":"2025-10-03T12:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.908690 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.908690 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:53:01 crc kubenswrapper[4578]: E1003 12:53:01.908940 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:53:01 crc kubenswrapper[4578]: E1003 12:53:01.908854 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.941383 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.941680 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.941793 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.941886 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:01 crc kubenswrapper[4578]: I1003 12:53:01.942158 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:01Z","lastTransitionTime":"2025-10-03T12:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.044454 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.044742 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.044822 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.044896 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.044959 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:02Z","lastTransitionTime":"2025-10-03T12:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.146871 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.146910 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.146920 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.146935 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.146945 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:02Z","lastTransitionTime":"2025-10-03T12:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.249338 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.249406 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.249419 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.249431 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.249439 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:02Z","lastTransitionTime":"2025-10-03T12:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.354659 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.355129 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.355221 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.355321 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.355401 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:02Z","lastTransitionTime":"2025-10-03T12:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.457164 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.457209 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.457222 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.457238 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.457250 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:02Z","lastTransitionTime":"2025-10-03T12:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.559778 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.559818 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.559828 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.559842 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.559851 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:02Z","lastTransitionTime":"2025-10-03T12:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.661831 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.662027 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.662125 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.662196 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.662265 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:02Z","lastTransitionTime":"2025-10-03T12:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.764021 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.764053 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.764061 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.764073 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.764081 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:02Z","lastTransitionTime":"2025-10-03T12:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.866955 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.866993 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.867005 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.867020 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.867033 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:02Z","lastTransitionTime":"2025-10-03T12:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.908519 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.908602 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:53:02 crc kubenswrapper[4578]: E1003 12:53:02.908811 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:53:02 crc kubenswrapper[4578]: E1003 12:53:02.908966 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.968847 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.968887 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.968898 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.968914 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:02 crc kubenswrapper[4578]: I1003 12:53:02.968928 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:02Z","lastTransitionTime":"2025-10-03T12:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.071447 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.071482 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.071493 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.071509 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.071520 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:03Z","lastTransitionTime":"2025-10-03T12:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.173675 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.173718 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.173730 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.173744 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.173755 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:03Z","lastTransitionTime":"2025-10-03T12:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.276114 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.276164 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.276175 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.276190 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.276205 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:03Z","lastTransitionTime":"2025-10-03T12:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.378550 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.378591 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.378600 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.378659 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.378670 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:03Z","lastTransitionTime":"2025-10-03T12:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.480585 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.480628 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.480662 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.480675 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.480684 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:03Z","lastTransitionTime":"2025-10-03T12:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.582707 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.582754 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.582766 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.582780 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.582791 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:03Z","lastTransitionTime":"2025-10-03T12:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.684492 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.684522 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.684530 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.684541 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.684549 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:03Z","lastTransitionTime":"2025-10-03T12:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.786950 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.787007 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.787015 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.787027 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.787036 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:03Z","lastTransitionTime":"2025-10-03T12:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.889015 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.889061 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.889076 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.889091 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.889101 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:03Z","lastTransitionTime":"2025-10-03T12:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.908336 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.908394 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:53:03 crc kubenswrapper[4578]: E1003 12:53:03.908463 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:53:03 crc kubenswrapper[4578]: E1003 12:53:03.908519 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.991331 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.991365 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.991375 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.991387 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:03 crc kubenswrapper[4578]: I1003 12:53:03.991395 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:03Z","lastTransitionTime":"2025-10-03T12:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.093253 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.093347 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.093360 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.093375 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.093384 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:04Z","lastTransitionTime":"2025-10-03T12:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.196225 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.196274 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.196286 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.196306 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.196318 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:04Z","lastTransitionTime":"2025-10-03T12:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.298081 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.298113 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.298121 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.298134 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.298144 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:04Z","lastTransitionTime":"2025-10-03T12:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.400416 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.400449 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.400457 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.400476 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.400491 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:04Z","lastTransitionTime":"2025-10-03T12:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.502992 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.503031 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.503042 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.503057 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.503068 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:04Z","lastTransitionTime":"2025-10-03T12:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.605682 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.605719 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.605730 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.605746 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.605758 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:04Z","lastTransitionTime":"2025-10-03T12:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.707754 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.707812 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.707823 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.707839 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.707849 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:04Z","lastTransitionTime":"2025-10-03T12:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.810185 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.810232 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.810240 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.810252 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.810260 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:04Z","lastTransitionTime":"2025-10-03T12:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.908283 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.908408 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:53:04 crc kubenswrapper[4578]: E1003 12:53:04.908544 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:53:04 crc kubenswrapper[4578]: E1003 12:53:04.908686 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:53:04 crc kubenswrapper[4578]: E1003 12:53:04.910549 4578 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.940932 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=20.940912597 podStartE2EDuration="20.940912597s" podCreationTimestamp="2025-10-03 12:52:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:04.931679579 +0000 UTC m=+120.730151763" watchObservedRunningTime="2025-10-03 12:53:04.940912597 +0000 UTC m=+120.739384781" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.954350 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=37.954321165 podStartE2EDuration="37.954321165s" podCreationTimestamp="2025-10-03 12:52:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:04.941359871 +0000 UTC m=+120.739832055" watchObservedRunningTime="2025-10-03 12:53:04.954321165 +0000 UTC m=+120.752793349" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.954815 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-c4jgn" podStartSLOduration=93.954809191 podStartE2EDuration="1m33.954809191s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:04.953447168 +0000 UTC m=+120.751919372" watchObservedRunningTime="2025-10-03 12:53:04.954809191 +0000 UTC m=+120.753281375" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.966115 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podStartSLOduration=93.966098814 podStartE2EDuration="1m33.966098814s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:04.965967229 +0000 UTC m=+120.764439403" watchObservedRunningTime="2025-10-03 12:53:04.966098814 +0000 UTC m=+120.764570998" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.987951 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-rh9f2" podStartSLOduration=94.987930385 podStartE2EDuration="1m34.987930385s" podCreationTimestamp="2025-10-03 12:51:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:04.977642983 +0000 UTC m=+120.776115167" watchObservedRunningTime="2025-10-03 12:53:04.987930385 +0000 UTC m=+120.786402559" Oct 03 12:53:04 crc kubenswrapper[4578]: I1003 12:53:04.988269 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=67.988263954 podStartE2EDuration="1m7.988263954s" podCreationTimestamp="2025-10-03 12:51:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:04.98780122 +0000 UTC m=+120.786273404" watchObservedRunningTime="2025-10-03 12:53:04.988263954 +0000 UTC m=+120.786736138" Oct 03 12:53:05 crc kubenswrapper[4578]: E1003 12:53:05.013569 4578 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 03 12:53:05 crc kubenswrapper[4578]: I1003 12:53:05.036494 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-p8d2v" podStartSLOduration=95.036475519 podStartE2EDuration="1m35.036475519s" podCreationTimestamp="2025-10-03 12:51:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:05.036127158 +0000 UTC m=+120.834599342" watchObservedRunningTime="2025-10-03 12:53:05.036475519 +0000 UTC m=+120.834947703" Oct 03 12:53:05 crc kubenswrapper[4578]: I1003 12:53:05.061967 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xc4j9" podStartSLOduration=94.061949683 podStartE2EDuration="1m34.061949683s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:05.050507826 +0000 UTC m=+120.848980010" watchObservedRunningTime="2025-10-03 12:53:05.061949683 +0000 UTC m=+120.860421867" Oct 03 12:53:05 crc kubenswrapper[4578]: I1003 12:53:05.092754 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-jjvmh" podStartSLOduration=94.092737953 podStartE2EDuration="1m34.092737953s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:05.092698663 +0000 UTC m=+120.891170847" watchObservedRunningTime="2025-10-03 12:53:05.092737953 +0000 UTC m=+120.891210137" Oct 03 12:53:05 crc kubenswrapper[4578]: I1003 12:53:05.108321 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=93.108304859 podStartE2EDuration="1m33.108304859s" podCreationTimestamp="2025-10-03 12:51:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:05.108103022 +0000 UTC m=+120.906575206" watchObservedRunningTime="2025-10-03 12:53:05.108304859 +0000 UTC m=+120.906777043" Oct 03 12:53:05 crc kubenswrapper[4578]: I1003 12:53:05.133032 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=22.13301432 podStartE2EDuration="22.13301432s" podCreationTimestamp="2025-10-03 12:52:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:05.130730379 +0000 UTC m=+120.929202573" watchObservedRunningTime="2025-10-03 12:53:05.13301432 +0000 UTC m=+120.931486504" Oct 03 12:53:05 crc kubenswrapper[4578]: I1003 12:53:05.908289 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:53:05 crc kubenswrapper[4578]: I1003 12:53:05.908300 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:53:05 crc kubenswrapper[4578]: E1003 12:53:05.908492 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:53:05 crc kubenswrapper[4578]: E1003 12:53:05.908409 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:53:06 crc kubenswrapper[4578]: I1003 12:53:06.909173 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:53:06 crc kubenswrapper[4578]: I1003 12:53:06.909173 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:06 crc kubenswrapper[4578]: E1003 12:53:06.909356 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:53:06 crc kubenswrapper[4578]: E1003 12:53:06.909438 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:53:07 crc kubenswrapper[4578]: I1003 12:53:07.908043 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:53:07 crc kubenswrapper[4578]: I1003 12:53:07.908075 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:53:07 crc kubenswrapper[4578]: E1003 12:53:07.908170 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:53:07 crc kubenswrapper[4578]: E1003 12:53:07.908277 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:53:08 crc kubenswrapper[4578]: I1003 12:53:08.532774 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-c4jgn_fed3968f-4bb0-4209-a377-16c26f9ea070/kube-multus/1.log" Oct 03 12:53:08 crc kubenswrapper[4578]: I1003 12:53:08.533224 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-c4jgn_fed3968f-4bb0-4209-a377-16c26f9ea070/kube-multus/0.log" Oct 03 12:53:08 crc kubenswrapper[4578]: I1003 12:53:08.533260 4578 generic.go:334] "Generic (PLEG): container finished" podID="fed3968f-4bb0-4209-a377-16c26f9ea070" containerID="650520feb6cfcf5c6aea04b4a990593eab6f2a2ddec00e2270105f25ee9a067d" exitCode=1 Oct 03 12:53:08 crc kubenswrapper[4578]: I1003 12:53:08.533289 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-c4jgn" event={"ID":"fed3968f-4bb0-4209-a377-16c26f9ea070","Type":"ContainerDied","Data":"650520feb6cfcf5c6aea04b4a990593eab6f2a2ddec00e2270105f25ee9a067d"} Oct 03 12:53:08 crc kubenswrapper[4578]: I1003 12:53:08.533323 4578 scope.go:117] "RemoveContainer" containerID="7afdac3fcec1ea3a5acdf1137760339b7c51e33e6b4e82ff4123392d50ccabc3" Oct 03 12:53:08 crc kubenswrapper[4578]: I1003 12:53:08.534160 4578 scope.go:117] "RemoveContainer" containerID="650520feb6cfcf5c6aea04b4a990593eab6f2a2ddec00e2270105f25ee9a067d" Oct 03 12:53:08 crc kubenswrapper[4578]: E1003 12:53:08.534310 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-c4jgn_openshift-multus(fed3968f-4bb0-4209-a377-16c26f9ea070)\"" pod="openshift-multus/multus-c4jgn" podUID="fed3968f-4bb0-4209-a377-16c26f9ea070" Oct 03 12:53:08 crc kubenswrapper[4578]: I1003 12:53:08.908520 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:53:08 crc kubenswrapper[4578]: I1003 12:53:08.908553 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:08 crc kubenswrapper[4578]: I1003 12:53:08.909100 4578 scope.go:117] "RemoveContainer" containerID="508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c" Oct 03 12:53:08 crc kubenswrapper[4578]: E1003 12:53:08.909209 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:53:08 crc kubenswrapper[4578]: E1003 12:53:08.909249 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8qvh9_openshift-ovn-kubernetes(ebaa96c5-be7c-4f05-819d-f37e1ceddac8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" Oct 03 12:53:08 crc kubenswrapper[4578]: E1003 12:53:08.909572 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.325284 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.325321 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.325332 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.325349 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.325361 4578 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-03T12:53:09Z","lastTransitionTime":"2025-10-03T12:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.371237 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm"] Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.371679 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.376158 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.376437 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.377165 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.377726 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.509422 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e87231c6-5146-4506-9a05-1cf2aab5e7ba-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-x6mjm\" (UID: \"e87231c6-5146-4506-9a05-1cf2aab5e7ba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.509480 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e87231c6-5146-4506-9a05-1cf2aab5e7ba-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-x6mjm\" (UID: \"e87231c6-5146-4506-9a05-1cf2aab5e7ba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.509515 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e87231c6-5146-4506-9a05-1cf2aab5e7ba-service-ca\") pod \"cluster-version-operator-5c965bbfc6-x6mjm\" (UID: \"e87231c6-5146-4506-9a05-1cf2aab5e7ba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.509548 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e87231c6-5146-4506-9a05-1cf2aab5e7ba-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-x6mjm\" (UID: \"e87231c6-5146-4506-9a05-1cf2aab5e7ba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.509605 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e87231c6-5146-4506-9a05-1cf2aab5e7ba-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-x6mjm\" (UID: \"e87231c6-5146-4506-9a05-1cf2aab5e7ba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.544673 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-c4jgn_fed3968f-4bb0-4209-a377-16c26f9ea070/kube-multus/1.log" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.610194 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e87231c6-5146-4506-9a05-1cf2aab5e7ba-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-x6mjm\" (UID: \"e87231c6-5146-4506-9a05-1cf2aab5e7ba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.610233 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e87231c6-5146-4506-9a05-1cf2aab5e7ba-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-x6mjm\" (UID: \"e87231c6-5146-4506-9a05-1cf2aab5e7ba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.610259 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e87231c6-5146-4506-9a05-1cf2aab5e7ba-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-x6mjm\" (UID: \"e87231c6-5146-4506-9a05-1cf2aab5e7ba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.610320 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e87231c6-5146-4506-9a05-1cf2aab5e7ba-service-ca\") pod \"cluster-version-operator-5c965bbfc6-x6mjm\" (UID: \"e87231c6-5146-4506-9a05-1cf2aab5e7ba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.610351 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e87231c6-5146-4506-9a05-1cf2aab5e7ba-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-x6mjm\" (UID: \"e87231c6-5146-4506-9a05-1cf2aab5e7ba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.610731 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e87231c6-5146-4506-9a05-1cf2aab5e7ba-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-x6mjm\" (UID: \"e87231c6-5146-4506-9a05-1cf2aab5e7ba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.610772 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e87231c6-5146-4506-9a05-1cf2aab5e7ba-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-x6mjm\" (UID: \"e87231c6-5146-4506-9a05-1cf2aab5e7ba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.611463 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e87231c6-5146-4506-9a05-1cf2aab5e7ba-service-ca\") pod \"cluster-version-operator-5c965bbfc6-x6mjm\" (UID: \"e87231c6-5146-4506-9a05-1cf2aab5e7ba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.620596 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e87231c6-5146-4506-9a05-1cf2aab5e7ba-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-x6mjm\" (UID: \"e87231c6-5146-4506-9a05-1cf2aab5e7ba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.626012 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e87231c6-5146-4506-9a05-1cf2aab5e7ba-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-x6mjm\" (UID: \"e87231c6-5146-4506-9a05-1cf2aab5e7ba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.687407 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.909171 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:53:09 crc kubenswrapper[4578]: I1003 12:53:09.909253 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:53:09 crc kubenswrapper[4578]: E1003 12:53:09.909430 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:53:09 crc kubenswrapper[4578]: E1003 12:53:09.909569 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:53:10 crc kubenswrapper[4578]: E1003 12:53:10.014165 4578 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 03 12:53:10 crc kubenswrapper[4578]: I1003 12:53:10.549911 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm" event={"ID":"e87231c6-5146-4506-9a05-1cf2aab5e7ba","Type":"ContainerStarted","Data":"bd25f24985b83df1ba57fb2303c2f89489fdae0e0378a93012f069c0b55ba1ba"} Oct 03 12:53:10 crc kubenswrapper[4578]: I1003 12:53:10.550071 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm" event={"ID":"e87231c6-5146-4506-9a05-1cf2aab5e7ba","Type":"ContainerStarted","Data":"080f323cecdaafd84019d65d2def2bfb0ef93ee5b227cb810ee639b6025379a5"} Oct 03 12:53:10 crc kubenswrapper[4578]: I1003 12:53:10.565035 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x6mjm" podStartSLOduration=100.565017396 podStartE2EDuration="1m40.565017396s" podCreationTimestamp="2025-10-03 12:51:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:10.564275703 +0000 UTC m=+126.362747897" watchObservedRunningTime="2025-10-03 12:53:10.565017396 +0000 UTC m=+126.363489580" Oct 03 12:53:10 crc kubenswrapper[4578]: I1003 12:53:10.908776 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:53:10 crc kubenswrapper[4578]: E1003 12:53:10.908900 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:53:10 crc kubenswrapper[4578]: I1003 12:53:10.909064 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:10 crc kubenswrapper[4578]: E1003 12:53:10.909120 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:53:11 crc kubenswrapper[4578]: I1003 12:53:11.907995 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:53:11 crc kubenswrapper[4578]: I1003 12:53:11.908038 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:53:11 crc kubenswrapper[4578]: E1003 12:53:11.908441 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:53:11 crc kubenswrapper[4578]: E1003 12:53:11.908514 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:53:12 crc kubenswrapper[4578]: I1003 12:53:12.908355 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:53:12 crc kubenswrapper[4578]: E1003 12:53:12.908529 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:53:12 crc kubenswrapper[4578]: I1003 12:53:12.908839 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:12 crc kubenswrapper[4578]: E1003 12:53:12.908905 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:53:13 crc kubenswrapper[4578]: I1003 12:53:13.908199 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:53:13 crc kubenswrapper[4578]: E1003 12:53:13.908348 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:53:13 crc kubenswrapper[4578]: I1003 12:53:13.908935 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:53:13 crc kubenswrapper[4578]: E1003 12:53:13.909147 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:53:14 crc kubenswrapper[4578]: I1003 12:53:14.908453 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:53:14 crc kubenswrapper[4578]: E1003 12:53:14.911215 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:53:14 crc kubenswrapper[4578]: I1003 12:53:14.911303 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:14 crc kubenswrapper[4578]: E1003 12:53:14.911387 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:53:15 crc kubenswrapper[4578]: E1003 12:53:15.016184 4578 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 03 12:53:15 crc kubenswrapper[4578]: I1003 12:53:15.909051 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:53:15 crc kubenswrapper[4578]: I1003 12:53:15.909051 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:53:15 crc kubenswrapper[4578]: E1003 12:53:15.909196 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:53:15 crc kubenswrapper[4578]: E1003 12:53:15.909276 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:53:16 crc kubenswrapper[4578]: I1003 12:53:16.908445 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:16 crc kubenswrapper[4578]: I1003 12:53:16.908527 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:53:16 crc kubenswrapper[4578]: E1003 12:53:16.908593 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:53:16 crc kubenswrapper[4578]: E1003 12:53:16.909051 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:53:17 crc kubenswrapper[4578]: I1003 12:53:17.908526 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:53:17 crc kubenswrapper[4578]: I1003 12:53:17.908598 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:53:17 crc kubenswrapper[4578]: E1003 12:53:17.908696 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:53:17 crc kubenswrapper[4578]: E1003 12:53:17.908805 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:53:18 crc kubenswrapper[4578]: I1003 12:53:18.908380 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:18 crc kubenswrapper[4578]: E1003 12:53:18.908576 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:53:18 crc kubenswrapper[4578]: I1003 12:53:18.908784 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:53:18 crc kubenswrapper[4578]: E1003 12:53:18.909086 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:53:19 crc kubenswrapper[4578]: I1003 12:53:19.909117 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:53:19 crc kubenswrapper[4578]: I1003 12:53:19.909151 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:53:19 crc kubenswrapper[4578]: E1003 12:53:19.909281 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:53:19 crc kubenswrapper[4578]: E1003 12:53:19.910681 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:53:20 crc kubenswrapper[4578]: E1003 12:53:20.017054 4578 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 03 12:53:20 crc kubenswrapper[4578]: I1003 12:53:20.908309 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:20 crc kubenswrapper[4578]: I1003 12:53:20.908377 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:53:20 crc kubenswrapper[4578]: E1003 12:53:20.908939 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:53:20 crc kubenswrapper[4578]: E1003 12:53:20.909099 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:53:21 crc kubenswrapper[4578]: I1003 12:53:21.908872 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:53:21 crc kubenswrapper[4578]: I1003 12:53:21.908959 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:53:21 crc kubenswrapper[4578]: E1003 12:53:21.909093 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:53:21 crc kubenswrapper[4578]: E1003 12:53:21.909290 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:53:21 crc kubenswrapper[4578]: I1003 12:53:21.909355 4578 scope.go:117] "RemoveContainer" containerID="650520feb6cfcf5c6aea04b4a990593eab6f2a2ddec00e2270105f25ee9a067d" Oct 03 12:53:22 crc kubenswrapper[4578]: I1003 12:53:22.587730 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-c4jgn_fed3968f-4bb0-4209-a377-16c26f9ea070/kube-multus/1.log" Oct 03 12:53:22 crc kubenswrapper[4578]: I1003 12:53:22.587794 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-c4jgn" event={"ID":"fed3968f-4bb0-4209-a377-16c26f9ea070","Type":"ContainerStarted","Data":"8863d0d89e062a4984f279b5b3f315ff53e3502946c377ef60a2481ab6294084"} Oct 03 12:53:22 crc kubenswrapper[4578]: I1003 12:53:22.908399 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:22 crc kubenswrapper[4578]: E1003 12:53:22.908572 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:53:22 crc kubenswrapper[4578]: I1003 12:53:22.909407 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:53:22 crc kubenswrapper[4578]: E1003 12:53:22.909609 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:53:22 crc kubenswrapper[4578]: I1003 12:53:22.909778 4578 scope.go:117] "RemoveContainer" containerID="508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c" Oct 03 12:53:23 crc kubenswrapper[4578]: I1003 12:53:23.591905 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8qvh9_ebaa96c5-be7c-4f05-819d-f37e1ceddac8/ovnkube-controller/3.log" Oct 03 12:53:23 crc kubenswrapper[4578]: I1003 12:53:23.594293 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerStarted","Data":"ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01"} Oct 03 12:53:23 crc kubenswrapper[4578]: I1003 12:53:23.594693 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:53:23 crc kubenswrapper[4578]: I1003 12:53:23.621056 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" podStartSLOduration=112.621037405 podStartE2EDuration="1m52.621037405s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:23.61966387 +0000 UTC m=+139.418136054" watchObservedRunningTime="2025-10-03 12:53:23.621037405 +0000 UTC m=+139.419509589" Oct 03 12:53:23 crc kubenswrapper[4578]: I1003 12:53:23.680125 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-q8986"] Oct 03 12:53:23 crc kubenswrapper[4578]: I1003 12:53:23.680248 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:53:23 crc kubenswrapper[4578]: E1003 12:53:23.680351 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:53:23 crc kubenswrapper[4578]: I1003 12:53:23.908268 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:53:23 crc kubenswrapper[4578]: I1003 12:53:23.908327 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:53:23 crc kubenswrapper[4578]: E1003 12:53:23.908379 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:53:23 crc kubenswrapper[4578]: E1003 12:53:23.908436 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:53:24 crc kubenswrapper[4578]: I1003 12:53:24.908457 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:24 crc kubenswrapper[4578]: E1003 12:53:24.909807 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:53:25 crc kubenswrapper[4578]: E1003 12:53:25.018657 4578 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 03 12:53:25 crc kubenswrapper[4578]: I1003 12:53:25.908270 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:53:25 crc kubenswrapper[4578]: I1003 12:53:25.908360 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:53:25 crc kubenswrapper[4578]: I1003 12:53:25.908371 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:53:25 crc kubenswrapper[4578]: E1003 12:53:25.908434 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:53:25 crc kubenswrapper[4578]: E1003 12:53:25.908516 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:53:25 crc kubenswrapper[4578]: E1003 12:53:25.908609 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:53:26 crc kubenswrapper[4578]: I1003 12:53:26.908625 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:26 crc kubenswrapper[4578]: E1003 12:53:26.908761 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:53:27 crc kubenswrapper[4578]: I1003 12:53:27.908854 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:53:27 crc kubenswrapper[4578]: I1003 12:53:27.908942 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:53:27 crc kubenswrapper[4578]: I1003 12:53:27.909018 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:53:27 crc kubenswrapper[4578]: E1003 12:53:27.909103 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:53:27 crc kubenswrapper[4578]: E1003 12:53:27.909156 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:53:27 crc kubenswrapper[4578]: E1003 12:53:27.909290 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:53:28 crc kubenswrapper[4578]: I1003 12:53:28.908912 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:28 crc kubenswrapper[4578]: E1003 12:53:28.909052 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 03 12:53:29 crc kubenswrapper[4578]: I1003 12:53:29.707451 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 12:53:29 crc kubenswrapper[4578]: I1003 12:53:29.908383 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:53:29 crc kubenswrapper[4578]: I1003 12:53:29.908423 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:53:29 crc kubenswrapper[4578]: I1003 12:53:29.908489 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:53:29 crc kubenswrapper[4578]: E1003 12:53:29.909025 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 03 12:53:29 crc kubenswrapper[4578]: E1003 12:53:29.908864 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-q8986" podUID="2971c7bb-e0f4-44a8-8f52-9ea8ba199772" Oct 03 12:53:29 crc kubenswrapper[4578]: E1003 12:53:29.909084 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 03 12:53:30 crc kubenswrapper[4578]: I1003 12:53:30.908320 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:30 crc kubenswrapper[4578]: I1003 12:53:30.910366 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 03 12:53:30 crc kubenswrapper[4578]: I1003 12:53:30.910503 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 03 12:53:31 crc kubenswrapper[4578]: I1003 12:53:31.908481 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:53:31 crc kubenswrapper[4578]: I1003 12:53:31.908517 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:53:31 crc kubenswrapper[4578]: I1003 12:53:31.908578 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:53:31 crc kubenswrapper[4578]: I1003 12:53:31.911215 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 03 12:53:31 crc kubenswrapper[4578]: I1003 12:53:31.911221 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 03 12:53:31 crc kubenswrapper[4578]: I1003 12:53:31.911218 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 03 12:53:31 crc kubenswrapper[4578]: I1003 12:53:31.912453 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 03 12:53:35 crc kubenswrapper[4578]: I1003 12:53:35.091331 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 12:53:35 crc kubenswrapper[4578]: I1003 12:53:35.091418 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 12:53:39 crc kubenswrapper[4578]: I1003 12:53:39.798034 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:39 crc kubenswrapper[4578]: E1003 12:53:39.798225 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:55:41.798209965 +0000 UTC m=+277.596682149 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:39 crc kubenswrapper[4578]: I1003 12:53:39.898768 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:53:39 crc kubenswrapper[4578]: I1003 12:53:39.898815 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:39 crc kubenswrapper[4578]: I1003 12:53:39.898851 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:39 crc kubenswrapper[4578]: I1003 12:53:39.898872 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:53:39 crc kubenswrapper[4578]: I1003 12:53:39.900159 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:39 crc kubenswrapper[4578]: I1003 12:53:39.903975 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:39 crc kubenswrapper[4578]: I1003 12:53:39.904190 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:53:39 crc kubenswrapper[4578]: I1003 12:53:39.904900 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:53:39 crc kubenswrapper[4578]: I1003 12:53:39.922264 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.022807 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.029047 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.051443 4578 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.087944 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hh7tq"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.088266 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.090560 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2cj4m"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.091046 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2cj4m" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.091865 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.092050 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7br6x"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.092340 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.092489 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cxghf"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.092858 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cxghf" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.093423 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.093785 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-wmkdd"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.094153 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.096201 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.101059 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.101270 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.102809 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-mt2gv"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.103295 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mt2gv" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.103430 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.112781 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.113102 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.113304 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.113480 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.113746 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.113920 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.114088 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.114800 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.114983 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.115194 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.115303 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.115339 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.115441 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.115464 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.115545 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.115600 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.115688 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.115752 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.115773 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.115859 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.115874 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.115963 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.116066 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.118809 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.120167 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.136783 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.138362 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.138789 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4s8fg"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.139041 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-4s8fg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.139321 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.140994 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.141758 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zsb7g"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.142449 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.142487 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.144061 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.144226 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.144389 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.144561 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.144741 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.149734 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.151926 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.153426 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.154803 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-ns26l"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.155228 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-bxmjt"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.155472 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hh7tq"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.155488 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b6rr5"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.155723 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-8htrd"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.155986 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.156020 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-d2rzv"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.156106 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8htrd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.156153 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-bxmjt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.156420 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-zsb7g" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.156497 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b6rr5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.158203 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-d2rzv" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.161081 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.162899 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.163085 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.163277 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.163405 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.163529 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.163671 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.163848 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.165100 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.165260 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.165410 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.165520 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.165613 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.165718 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2cj4m"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.165750 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5svf7"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.166192 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5svf7" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.166585 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cl28q"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.166694 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.166800 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.166881 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.167037 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.167427 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.169307 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.169649 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-5bv45"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.170023 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-5bv45" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.170112 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7kdcr"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.170191 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.170482 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.170597 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7kdcr" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.171557 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.173729 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4qsl8"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.174138 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8flns"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.186588 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xctz5"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.192093 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4qsl8" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.196038 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xctz5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.202587 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8flns" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.211745 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a11b49f-69d2-4d9d-a451-7b69b797da20-config\") pod \"console-operator-58897d9998-bxmjt\" (UID: \"9a11b49f-69d2-4d9d-a451-7b69b797da20\") " pod="openshift-console-operator/console-operator-58897d9998-bxmjt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.211797 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.211835 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9ffa547-6ab2-4884-8433-d80fff56b441-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5svf7\" (UID: \"e9ffa547-6ab2-4884-8433-d80fff56b441\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5svf7" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.211857 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/185f4f54-9f50-4017-a7fc-7c7821338442-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2cj4m\" (UID: \"185f4f54-9f50-4017-a7fc-7c7821338442\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2cj4m" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.211875 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ad129323-bfad-47a1-84a5-7b51025f5d0d-image-import-ca\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.211893 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bec2b13f-9bf4-4378-8a9c-56e61397071a-encryption-config\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.211912 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9mzs\" (UniqueName: \"kubernetes.io/projected/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-kube-api-access-h9mzs\") pod \"route-controller-manager-6576b87f9c-t9q7n\" (UID: \"5a1d0b1c-9bae-483f-886d-bf4acffee7d9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.211931 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6c98467a-2f3b-4329-8f21-9492b087b5d8-metrics-tls\") pod \"dns-operator-744455d44c-5bv45\" (UID: \"6c98467a-2f3b-4329-8f21-9492b087b5d8\") " pod="openshift-dns-operator/dns-operator-744455d44c-5bv45" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.211948 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.211966 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dea78529-724d-461e-a64c-ed4858ff983c-config\") pod \"authentication-operator-69f744f599-4s8fg\" (UID: \"dea78529-724d-461e-a64c-ed4858ff983c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4s8fg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.211983 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2d27f0e8-6b63-4128-b886-7f0018a6a18b-etcd-client\") pod \"etcd-operator-b45778765-cl28q\" (UID: \"2d27f0e8-6b63-4128-b886-7f0018a6a18b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212004 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vj4m\" (UniqueName: \"kubernetes.io/projected/185f4f54-9f50-4017-a7fc-7c7821338442-kube-api-access-4vj4m\") pod \"openshift-apiserver-operator-796bbdcf4f-2cj4m\" (UID: \"185f4f54-9f50-4017-a7fc-7c7821338442\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2cj4m" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212021 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-client-ca\") pod \"route-controller-manager-6576b87f9c-t9q7n\" (UID: \"5a1d0b1c-9bae-483f-886d-bf4acffee7d9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212040 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-console-config\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212060 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5g25\" (UniqueName: \"kubernetes.io/projected/2d27f0e8-6b63-4128-b886-7f0018a6a18b-kube-api-access-n5g25\") pod \"etcd-operator-b45778765-cl28q\" (UID: \"2d27f0e8-6b63-4128-b886-7f0018a6a18b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212097 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d27f0e8-6b63-4128-b886-7f0018a6a18b-config\") pod \"etcd-operator-b45778765-cl28q\" (UID: \"2d27f0e8-6b63-4128-b886-7f0018a6a18b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212116 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212133 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2ee8da6-43e0-496d-aea1-2c7849305419-serving-cert\") pod \"controller-manager-879f6c89f-hh7tq\" (UID: \"b2ee8da6-43e0-496d-aea1-2c7849305419\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212155 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5vnm\" (UniqueName: \"kubernetes.io/projected/6c98467a-2f3b-4329-8f21-9492b087b5d8-kube-api-access-n5vnm\") pod \"dns-operator-744455d44c-5bv45\" (UID: \"6c98467a-2f3b-4329-8f21-9492b087b5d8\") " pod="openshift-dns-operator/dns-operator-744455d44c-5bv45" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212173 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/185f4f54-9f50-4017-a7fc-7c7821338442-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2cj4m\" (UID: \"185f4f54-9f50-4017-a7fc-7c7821338442\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2cj4m" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212193 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad129323-bfad-47a1-84a5-7b51025f5d0d-config\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212209 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c48e4a1e-a25a-49d8-aa80-bd1e7425ae33-available-featuregates\") pod \"openshift-config-operator-7777fb866f-8htrd\" (UID: \"c48e4a1e-a25a-49d8-aa80-bd1e7425ae33\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8htrd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212228 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bec2b13f-9bf4-4378-8a9c-56e61397071a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212246 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bec2b13f-9bf4-4378-8a9c-56e61397071a-serving-cert\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212264 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ad129323-bfad-47a1-84a5-7b51025f5d0d-node-pullsecrets\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212282 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ad129323-bfad-47a1-84a5-7b51025f5d0d-audit-dir\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212297 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a11b49f-69d2-4d9d-a451-7b69b797da20-serving-cert\") pod \"console-operator-58897d9998-bxmjt\" (UID: \"9a11b49f-69d2-4d9d-a451-7b69b797da20\") " pod="openshift-console-operator/console-operator-58897d9998-bxmjt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212316 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2ee8da6-43e0-496d-aea1-2c7849305419-config\") pod \"controller-manager-879f6c89f-hh7tq\" (UID: \"b2ee8da6-43e0-496d-aea1-2c7849305419\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212335 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-468cd\" (UniqueName: \"kubernetes.io/projected/bec2b13f-9bf4-4378-8a9c-56e61397071a-kube-api-access-468cd\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212352 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dea78529-724d-461e-a64c-ed4858ff983c-serving-cert\") pod \"authentication-operator-69f744f599-4s8fg\" (UID: \"dea78529-724d-461e-a64c-ed4858ff983c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4s8fg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212368 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-audit-dir\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212388 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212408 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bec2b13f-9bf4-4378-8a9c-56e61397071a-audit-dir\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212426 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpc5w\" (UniqueName: \"kubernetes.io/projected/6b207ee8-d007-408b-8519-b6fa02401fd0-kube-api-access-jpc5w\") pod \"openshift-controller-manager-operator-756b6f6bc6-b6rr5\" (UID: \"6b207ee8-d007-408b-8519-b6fa02401fd0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b6rr5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212444 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c48e4a1e-a25a-49d8-aa80-bd1e7425ae33-serving-cert\") pod \"openshift-config-operator-7777fb866f-8htrd\" (UID: \"c48e4a1e-a25a-49d8-aa80-bd1e7425ae33\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8htrd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212474 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-serving-cert\") pod \"route-controller-manager-6576b87f9c-t9q7n\" (UID: \"5a1d0b1c-9bae-483f-886d-bf4acffee7d9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212503 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-audit-policies\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212520 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-config\") pod \"route-controller-manager-6576b87f9c-t9q7n\" (UID: \"5a1d0b1c-9bae-483f-886d-bf4acffee7d9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212539 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1a194512-24b8-43b7-89f9-db084d65ee0a-auth-proxy-config\") pod \"machine-approver-56656f9798-mt2gv\" (UID: \"1a194512-24b8-43b7-89f9-db084d65ee0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mt2gv" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212554 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a194512-24b8-43b7-89f9-db084d65ee0a-config\") pod \"machine-approver-56656f9798-mt2gv\" (UID: \"1a194512-24b8-43b7-89f9-db084d65ee0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mt2gv" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212573 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjg95\" (UniqueName: \"kubernetes.io/projected/1a194512-24b8-43b7-89f9-db084d65ee0a-kube-api-access-wjg95\") pod \"machine-approver-56656f9798-mt2gv\" (UID: \"1a194512-24b8-43b7-89f9-db084d65ee0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mt2gv" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212591 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cnqw\" (UniqueName: \"kubernetes.io/projected/de78c746-e145-48ba-8601-a3ff2c6b2b62-kube-api-access-9cnqw\") pod \"downloads-7954f5f757-d2rzv\" (UID: \"de78c746-e145-48ba-8601-a3ff2c6b2b62\") " pod="openshift-console/downloads-7954f5f757-d2rzv" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212614 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ad129323-bfad-47a1-84a5-7b51025f5d0d-etcd-serving-ca\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212649 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/2f691127-6990-46bd-afa7-d6f3f7eaa329-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zsb7g\" (UID: \"2f691127-6990-46bd-afa7-d6f3f7eaa329\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zsb7g" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212670 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbxfv\" (UniqueName: \"kubernetes.io/projected/bc81f2fa-5ee7-4034-81d3-a11639762ac9-kube-api-access-vbxfv\") pod \"cluster-samples-operator-665b6dd947-cxghf\" (UID: \"bc81f2fa-5ee7-4034-81d3-a11639762ac9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cxghf" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212718 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6j4z\" (UniqueName: \"kubernetes.io/projected/dea78529-724d-461e-a64c-ed4858ff983c-kube-api-access-m6j4z\") pod \"authentication-operator-69f744f599-4s8fg\" (UID: \"dea78529-724d-461e-a64c-ed4858ff983c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4s8fg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212737 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212755 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-console-oauth-config\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212772 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d564c\" (UniqueName: \"kubernetes.io/projected/e9ffa547-6ab2-4884-8433-d80fff56b441-kube-api-access-d564c\") pod \"cluster-image-registry-operator-dc59b4c8b-5svf7\" (UID: \"e9ffa547-6ab2-4884-8433-d80fff56b441\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5svf7" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212789 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ad129323-bfad-47a1-84a5-7b51025f5d0d-etcd-client\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212807 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg4wd\" (UniqueName: \"kubernetes.io/projected/ad129323-bfad-47a1-84a5-7b51025f5d0d-kube-api-access-jg4wd\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212825 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/1a194512-24b8-43b7-89f9-db084d65ee0a-machine-approver-tls\") pod \"machine-approver-56656f9798-mt2gv\" (UID: \"1a194512-24b8-43b7-89f9-db084d65ee0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mt2gv" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212841 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b207ee8-d007-408b-8519-b6fa02401fd0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-b6rr5\" (UID: \"6b207ee8-d007-408b-8519-b6fa02401fd0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b6rr5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212861 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h79j6\" (UniqueName: \"kubernetes.io/projected/c48e4a1e-a25a-49d8-aa80-bd1e7425ae33-kube-api-access-h79j6\") pod \"openshift-config-operator-7777fb866f-8htrd\" (UID: \"c48e4a1e-a25a-49d8-aa80-bd1e7425ae33\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8htrd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212880 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212898 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dea78529-724d-461e-a64c-ed4858ff983c-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4s8fg\" (UID: \"dea78529-724d-461e-a64c-ed4858ff983c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4s8fg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212917 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e9ffa547-6ab2-4884-8433-d80fff56b441-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5svf7\" (UID: \"e9ffa547-6ab2-4884-8433-d80fff56b441\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5svf7" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212933 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad129323-bfad-47a1-84a5-7b51025f5d0d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212951 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbjhc\" (UniqueName: \"kubernetes.io/projected/2f691127-6990-46bd-afa7-d6f3f7eaa329-kube-api-access-kbjhc\") pod \"machine-api-operator-5694c8668f-zsb7g\" (UID: \"2f691127-6990-46bd-afa7-d6f3f7eaa329\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zsb7g" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212968 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ad129323-bfad-47a1-84a5-7b51025f5d0d-audit\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.212987 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2f691127-6990-46bd-afa7-d6f3f7eaa329-images\") pod \"machine-api-operator-5694c8668f-zsb7g\" (UID: \"2f691127-6990-46bd-afa7-d6f3f7eaa329\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zsb7g" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213002 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ad129323-bfad-47a1-84a5-7b51025f5d0d-encryption-config\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213019 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9a11b49f-69d2-4d9d-a451-7b69b797da20-trusted-ca\") pod \"console-operator-58897d9998-bxmjt\" (UID: \"9a11b49f-69d2-4d9d-a451-7b69b797da20\") " pod="openshift-console-operator/console-operator-58897d9998-bxmjt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213036 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213056 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bec2b13f-9bf4-4378-8a9c-56e61397071a-etcd-client\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213070 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-service-ca\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213088 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjfxr\" (UniqueName: \"kubernetes.io/projected/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-kube-api-access-bjfxr\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213119 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-trusted-ca-bundle\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213136 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkdx2\" (UniqueName: \"kubernetes.io/projected/9a11b49f-69d2-4d9d-a451-7b69b797da20-kube-api-access-kkdx2\") pod \"console-operator-58897d9998-bxmjt\" (UID: \"9a11b49f-69d2-4d9d-a451-7b69b797da20\") " pod="openshift-console-operator/console-operator-58897d9998-bxmjt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213154 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkv6n\" (UniqueName: \"kubernetes.io/projected/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-kube-api-access-nkv6n\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213173 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bec2b13f-9bf4-4378-8a9c-56e61397071a-audit-policies\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213192 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2d27f0e8-6b63-4128-b886-7f0018a6a18b-etcd-service-ca\") pod \"etcd-operator-b45778765-cl28q\" (UID: \"2d27f0e8-6b63-4128-b886-7f0018a6a18b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213212 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g28ws\" (UniqueName: \"kubernetes.io/projected/b2ee8da6-43e0-496d-aea1-2c7849305419-kube-api-access-g28ws\") pod \"controller-manager-879f6c89f-hh7tq\" (UID: \"b2ee8da6-43e0-496d-aea1-2c7849305419\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213231 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/bc81f2fa-5ee7-4034-81d3-a11639762ac9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-cxghf\" (UID: \"bc81f2fa-5ee7-4034-81d3-a11639762ac9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cxghf" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213249 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bec2b13f-9bf4-4378-8a9c-56e61397071a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213267 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-oauth-serving-cert\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213284 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b2ee8da6-43e0-496d-aea1-2c7849305419-client-ca\") pod \"controller-manager-879f6c89f-hh7tq\" (UID: \"b2ee8da6-43e0-496d-aea1-2c7849305419\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213301 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213319 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2d27f0e8-6b63-4128-b886-7f0018a6a18b-etcd-ca\") pod \"etcd-operator-b45778765-cl28q\" (UID: \"2d27f0e8-6b63-4128-b886-7f0018a6a18b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213337 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213355 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213374 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b2ee8da6-43e0-496d-aea1-2c7849305419-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hh7tq\" (UID: \"b2ee8da6-43e0-496d-aea1-2c7849305419\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213406 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b207ee8-d007-408b-8519-b6fa02401fd0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-b6rr5\" (UID: \"6b207ee8-d007-408b-8519-b6fa02401fd0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b6rr5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213437 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d27f0e8-6b63-4128-b886-7f0018a6a18b-serving-cert\") pod \"etcd-operator-b45778765-cl28q\" (UID: \"2d27f0e8-6b63-4128-b886-7f0018a6a18b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213463 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad129323-bfad-47a1-84a5-7b51025f5d0d-serving-cert\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.213494 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f691127-6990-46bd-afa7-d6f3f7eaa329-config\") pod \"machine-api-operator-5694c8668f-zsb7g\" (UID: \"2f691127-6990-46bd-afa7-d6f3f7eaa329\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zsb7g" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.215710 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.215745 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dea78529-724d-461e-a64c-ed4858ff983c-service-ca-bundle\") pod \"authentication-operator-69f744f599-4s8fg\" (UID: \"dea78529-724d-461e-a64c-ed4858ff983c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4s8fg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.215771 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-console-serving-cert\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.215792 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9ffa547-6ab2-4884-8433-d80fff56b441-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5svf7\" (UID: \"e9ffa547-6ab2-4884-8433-d80fff56b441\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5svf7" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.218998 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.228470 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.228804 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.229004 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.229039 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.230960 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.231108 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.231222 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.231403 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.231569 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.232879 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.232975 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.237552 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.250509 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.251184 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.251554 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.251728 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.251790 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.251850 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.252028 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.252434 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.252602 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.252738 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.252881 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.253129 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.253215 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.253938 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.254146 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.254278 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.254421 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5xxhk"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.254446 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.254589 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.254793 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-m68m5"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.255207 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gmbjm"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.255893 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.258407 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.258842 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.262248 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hpc5d"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.262533 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.262689 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hpc5d" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.262891 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.262911 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.262941 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m68m5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.262962 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gmbjm" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.262604 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sqbkd"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.263747 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-q8kdg"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.263879 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.264010 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-r9hlj"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.264170 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.264655 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-r9hlj" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.264315 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.264829 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.264602 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-77kj8"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.267073 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.267415 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-l2z5f"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.272457 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ns26l"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.272620 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.267639 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.264503 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.273276 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-l2z5f" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.267624 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-77kj8" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.264992 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.276442 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.281157 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.265153 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.282520 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.266794 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.267958 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.268190 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.286796 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7br6x"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.295328 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.301225 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.301815 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.335958 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340169 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c48e4a1e-a25a-49d8-aa80-bd1e7425ae33-available-featuregates\") pod \"openshift-config-operator-7777fb866f-8htrd\" (UID: \"c48e4a1e-a25a-49d8-aa80-bd1e7425ae33\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8htrd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340216 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-m68m5\" (UID: \"fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m68m5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340241 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bec2b13f-9bf4-4378-8a9c-56e61397071a-serving-cert\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340262 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340282 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-468cd\" (UniqueName: \"kubernetes.io/projected/bec2b13f-9bf4-4378-8a9c-56e61397071a-kube-api-access-468cd\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340302 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dea78529-724d-461e-a64c-ed4858ff983c-serving-cert\") pod \"authentication-operator-69f744f599-4s8fg\" (UID: \"dea78529-724d-461e-a64c-ed4858ff983c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4s8fg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340330 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2ee8da6-43e0-496d-aea1-2c7849305419-config\") pod \"controller-manager-879f6c89f-hh7tq\" (UID: \"b2ee8da6-43e0-496d-aea1-2c7849305419\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340406 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-serving-cert\") pod \"route-controller-manager-6576b87f9c-t9q7n\" (UID: \"5a1d0b1c-9bae-483f-886d-bf4acffee7d9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340431 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1a194512-24b8-43b7-89f9-db084d65ee0a-auth-proxy-config\") pod \"machine-approver-56656f9798-mt2gv\" (UID: \"1a194512-24b8-43b7-89f9-db084d65ee0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mt2gv" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340460 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a194512-24b8-43b7-89f9-db084d65ee0a-config\") pod \"machine-approver-56656f9798-mt2gv\" (UID: \"1a194512-24b8-43b7-89f9-db084d65ee0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mt2gv" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340486 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjg95\" (UniqueName: \"kubernetes.io/projected/1a194512-24b8-43b7-89f9-db084d65ee0a-kube-api-access-wjg95\") pod \"machine-approver-56656f9798-mt2gv\" (UID: \"1a194512-24b8-43b7-89f9-db084d65ee0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mt2gv" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340508 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-audit-policies\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340530 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjddt\" (UniqueName: \"kubernetes.io/projected/59bf594f-62b6-4a2f-a50d-30b5d8357133-kube-api-access-fjddt\") pod \"router-default-5444994796-q8kdg\" (UID: \"59bf594f-62b6-4a2f-a50d-30b5d8357133\") " pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340555 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340554 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbxfv\" (UniqueName: \"kubernetes.io/projected/bc81f2fa-5ee7-4034-81d3-a11639762ac9-kube-api-access-vbxfv\") pod \"cluster-samples-operator-665b6dd947-cxghf\" (UID: \"bc81f2fa-5ee7-4034-81d3-a11639762ac9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cxghf" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340665 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cnqw\" (UniqueName: \"kubernetes.io/projected/de78c746-e145-48ba-8601-a3ff2c6b2b62-kube-api-access-9cnqw\") pod \"downloads-7954f5f757-d2rzv\" (UID: \"de78c746-e145-48ba-8601-a3ff2c6b2b62\") " pod="openshift-console/downloads-7954f5f757-d2rzv" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340689 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/2f691127-6990-46bd-afa7-d6f3f7eaa329-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zsb7g\" (UID: \"2f691127-6990-46bd-afa7-d6f3f7eaa329\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zsb7g" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340715 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6j4z\" (UniqueName: \"kubernetes.io/projected/dea78529-724d-461e-a64c-ed4858ff983c-kube-api-access-m6j4z\") pod \"authentication-operator-69f744f599-4s8fg\" (UID: \"dea78529-724d-461e-a64c-ed4858ff983c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4s8fg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340737 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340763 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a3d0383-95ba-457c-af8c-0eebe774e2ea-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gmbjm\" (UID: \"1a3d0383-95ba-457c-af8c-0eebe774e2ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gmbjm" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340838 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/959dff27-2ebe-417a-b5be-3e814dff3797-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-4qsl8\" (UID: \"959dff27-2ebe-417a-b5be-3e814dff3797\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4qsl8" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340967 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/09d82c69-039a-4e2a-bcd7-dad26f5c8486-proxy-tls\") pod \"machine-config-operator-74547568cd-8flns\" (UID: \"09d82c69-039a-4e2a-bcd7-dad26f5c8486\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8flns" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.340993 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc-proxy-tls\") pod \"machine-config-controller-84d6567774-m68m5\" (UID: \"fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m68m5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341011 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e9ffa547-6ab2-4884-8433-d80fff56b441-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5svf7\" (UID: \"e9ffa547-6ab2-4884-8433-d80fff56b441\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5svf7" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341033 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbjhc\" (UniqueName: \"kubernetes.io/projected/2f691127-6990-46bd-afa7-d6f3f7eaa329-kube-api-access-kbjhc\") pod \"machine-api-operator-5694c8668f-zsb7g\" (UID: \"2f691127-6990-46bd-afa7-d6f3f7eaa329\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zsb7g" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341053 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdtwh\" (UniqueName: \"kubernetes.io/projected/fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc-kube-api-access-rdtwh\") pod \"machine-config-controller-84d6567774-m68m5\" (UID: \"fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m68m5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341075 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dea78529-724d-461e-a64c-ed4858ff983c-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4s8fg\" (UID: \"dea78529-724d-461e-a64c-ed4858ff983c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4s8fg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341098 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7wfh\" (UniqueName: \"kubernetes.io/projected/7d9d9160-dd25-4640-a84b-ee9558be3b88-kube-api-access-j7wfh\") pod \"migrator-59844c95c7-77kj8\" (UID: \"7d9d9160-dd25-4640-a84b-ee9558be3b88\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-77kj8" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341211 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ad129323-bfad-47a1-84a5-7b51025f5d0d-encryption-config\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341233 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9a11b49f-69d2-4d9d-a451-7b69b797da20-trusted-ca\") pod \"console-operator-58897d9998-bxmjt\" (UID: \"9a11b49f-69d2-4d9d-a451-7b69b797da20\") " pod="openshift-console-operator/console-operator-58897d9998-bxmjt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341257 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bec2b13f-9bf4-4378-8a9c-56e61397071a-etcd-client\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341282 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-service-ca\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341302 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjfxr\" (UniqueName: \"kubernetes.io/projected/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-kube-api-access-bjfxr\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341333 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-trusted-ca-bundle\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341352 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bec2b13f-9bf4-4378-8a9c-56e61397071a-audit-policies\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341453 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xx2d\" (UniqueName: \"kubernetes.io/projected/ee3b2702-a354-4921-a66d-650bedd0c1b2-kube-api-access-8xx2d\") pod \"catalog-operator-68c6474976-l9fbz\" (UID: \"ee3b2702-a354-4921-a66d-650bedd0c1b2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341476 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-oauth-serving-cert\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341494 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b2ee8da6-43e0-496d-aea1-2c7849305419-client-ca\") pod \"controller-manager-879f6c89f-hh7tq\" (UID: \"b2ee8da6-43e0-496d-aea1-2c7849305419\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341509 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341537 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b2ee8da6-43e0-496d-aea1-2c7849305419-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hh7tq\" (UID: \"b2ee8da6-43e0-496d-aea1-2c7849305419\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341559 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b207ee8-d007-408b-8519-b6fa02401fd0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-b6rr5\" (UID: \"6b207ee8-d007-408b-8519-b6fa02401fd0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b6rr5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341581 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f691127-6990-46bd-afa7-d6f3f7eaa329-config\") pod \"machine-api-operator-5694c8668f-zsb7g\" (UID: \"2f691127-6990-46bd-afa7-d6f3f7eaa329\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zsb7g" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341682 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dea78529-724d-461e-a64c-ed4858ff983c-service-ca-bundle\") pod \"authentication-operator-69f744f599-4s8fg\" (UID: \"dea78529-724d-461e-a64c-ed4858ff983c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4s8fg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341699 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ee3b2702-a354-4921-a66d-650bedd0c1b2-srv-cert\") pod \"catalog-operator-68c6474976-l9fbz\" (UID: \"ee3b2702-a354-4921-a66d-650bedd0c1b2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341723 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4ef05574-4924-4a73-b0ba-0e632c96410a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sqbkd\" (UID: \"4ef05574-4924-4a73-b0ba-0e632c96410a\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341745 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9ffa547-6ab2-4884-8433-d80fff56b441-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5svf7\" (UID: \"e9ffa547-6ab2-4884-8433-d80fff56b441\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5svf7" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341763 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/959dff27-2ebe-417a-b5be-3e814dff3797-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-4qsl8\" (UID: \"959dff27-2ebe-417a-b5be-3e814dff3797\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4qsl8" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341781 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9ffa547-6ab2-4884-8433-d80fff56b441-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5svf7\" (UID: \"e9ffa547-6ab2-4884-8433-d80fff56b441\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5svf7" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341854 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a11b49f-69d2-4d9d-a451-7b69b797da20-config\") pod \"console-operator-58897d9998-bxmjt\" (UID: \"9a11b49f-69d2-4d9d-a451-7b69b797da20\") " pod="openshift-console-operator/console-operator-58897d9998-bxmjt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341878 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341905 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bec2b13f-9bf4-4378-8a9c-56e61397071a-encryption-config\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341932 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9mzs\" (UniqueName: \"kubernetes.io/projected/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-kube-api-access-h9mzs\") pod \"route-controller-manager-6576b87f9c-t9q7n\" (UID: \"5a1d0b1c-9bae-483f-886d-bf4acffee7d9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341950 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/09d82c69-039a-4e2a-bcd7-dad26f5c8486-images\") pod \"machine-config-operator-74547568cd-8flns\" (UID: \"09d82c69-039a-4e2a-bcd7-dad26f5c8486\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8flns" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341970 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dea78529-724d-461e-a64c-ed4858ff983c-config\") pod \"authentication-operator-69f744f599-4s8fg\" (UID: \"dea78529-724d-461e-a64c-ed4858ff983c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4s8fg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.341989 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-client-ca\") pod \"route-controller-manager-6576b87f9c-t9q7n\" (UID: \"5a1d0b1c-9bae-483f-886d-bf4acffee7d9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.342066 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-console-config\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.342091 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db43e12f-df81-44cf-9220-2f75d5ecc70b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xctz5\" (UID: \"db43e12f-df81-44cf-9220-2f75d5ecc70b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xctz5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.342117 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c5bd\" (UniqueName: \"kubernetes.io/projected/f78b5fb4-0a61-4222-928f-2383bac97358-kube-api-access-7c5bd\") pod \"multus-admission-controller-857f4d67dd-r9hlj\" (UID: \"f78b5fb4-0a61-4222-928f-2383bac97358\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-r9hlj" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.342138 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/92275d4e-f157-49a7-b292-2cd46bc23c11-signing-cabundle\") pod \"service-ca-9c57cc56f-l2z5f\" (UID: \"92275d4e-f157-49a7-b292-2cd46bc23c11\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2z5f" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.342157 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7wjj\" (UniqueName: \"kubernetes.io/projected/6f3f593c-f784-4f1c-8fb3-016275c38649-kube-api-access-n7wjj\") pod \"control-plane-machine-set-operator-78cbb6b69f-hpc5d\" (UID: \"6f3f593c-f784-4f1c-8fb3-016275c38649\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hpc5d" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.342179 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5581f00e-5f39-454d-843d-9dbf6d2cbb0b-metrics-tls\") pod \"ingress-operator-5b745b69d9-s2z8b\" (UID: \"5581f00e-5f39-454d-843d-9dbf6d2cbb0b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.342199 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.342282 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/185f4f54-9f50-4017-a7fc-7c7821338442-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2cj4m\" (UID: \"185f4f54-9f50-4017-a7fc-7c7821338442\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2cj4m" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.342312 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad129323-bfad-47a1-84a5-7b51025f5d0d-config\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.342342 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bec2b13f-9bf4-4378-8a9c-56e61397071a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.342365 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-audit-dir\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.342394 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/59bf594f-62b6-4a2f-a50d-30b5d8357133-stats-auth\") pod \"router-default-5444994796-q8kdg\" (UID: \"59bf594f-62b6-4a2f-a50d-30b5d8357133\") " pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.342423 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ad129323-bfad-47a1-84a5-7b51025f5d0d-node-pullsecrets\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.342461 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ad129323-bfad-47a1-84a5-7b51025f5d0d-audit-dir\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.342542 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a11b49f-69d2-4d9d-a451-7b69b797da20-serving-cert\") pod \"console-operator-58897d9998-bxmjt\" (UID: \"9a11b49f-69d2-4d9d-a451-7b69b797da20\") " pod="openshift-console-operator/console-operator-58897d9998-bxmjt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.342566 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c48e4a1e-a25a-49d8-aa80-bd1e7425ae33-serving-cert\") pod \"openshift-config-operator-7777fb866f-8htrd\" (UID: \"c48e4a1e-a25a-49d8-aa80-bd1e7425ae33\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8htrd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.342574 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-t9rgf"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.342592 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bec2b13f-9bf4-4378-8a9c-56e61397071a-audit-dir\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.343265 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.344086 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b2ee8da6-43e0-496d-aea1-2c7849305419-client-ca\") pod \"controller-manager-879f6c89f-hh7tq\" (UID: \"b2ee8da6-43e0-496d-aea1-2c7849305419\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.344818 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.347136 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b2ee8da6-43e0-496d-aea1-2c7849305419-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hh7tq\" (UID: \"b2ee8da6-43e0-496d-aea1-2c7849305419\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.348665 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.348882 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-oauth-serving-cert\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.350726 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c48e4a1e-a25a-49d8-aa80-bd1e7425ae33-available-featuregates\") pod \"openshift-config-operator-7777fb866f-8htrd\" (UID: \"c48e4a1e-a25a-49d8-aa80-bd1e7425ae33\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8htrd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.351793 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b207ee8-d007-408b-8519-b6fa02401fd0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-b6rr5\" (UID: \"6b207ee8-d007-408b-8519-b6fa02401fd0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b6rr5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.352511 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f691127-6990-46bd-afa7-d6f3f7eaa329-config\") pod \"machine-api-operator-5694c8668f-zsb7g\" (UID: \"2f691127-6990-46bd-afa7-d6f3f7eaa329\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zsb7g" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.353141 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dea78529-724d-461e-a64c-ed4858ff983c-service-ca-bundle\") pod \"authentication-operator-69f744f599-4s8fg\" (UID: \"dea78529-724d-461e-a64c-ed4858ff983c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4s8fg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.354209 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9ffa547-6ab2-4884-8433-d80fff56b441-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5svf7\" (UID: \"e9ffa547-6ab2-4884-8433-d80fff56b441\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5svf7" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.357246 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a11b49f-69d2-4d9d-a451-7b69b797da20-config\") pod \"console-operator-58897d9998-bxmjt\" (UID: \"9a11b49f-69d2-4d9d-a451-7b69b797da20\") " pod="openshift-console-operator/console-operator-58897d9998-bxmjt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.358746 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.359415 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.342624 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpc5w\" (UniqueName: \"kubernetes.io/projected/6b207ee8-d007-408b-8519-b6fa02401fd0-kube-api-access-jpc5w\") pod \"openshift-controller-manager-operator-756b6f6bc6-b6rr5\" (UID: \"6b207ee8-d007-408b-8519-b6fa02401fd0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b6rr5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.360084 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.360385 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-config\") pod \"route-controller-manager-6576b87f9c-t9q7n\" (UID: \"5a1d0b1c-9bae-483f-886d-bf4acffee7d9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.360418 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjmr7\" (UniqueName: \"kubernetes.io/projected/09d82c69-039a-4e2a-bcd7-dad26f5c8486-kube-api-access-kjmr7\") pod \"machine-config-operator-74547568cd-8flns\" (UID: \"09d82c69-039a-4e2a-bcd7-dad26f5c8486\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8flns" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.360448 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ad129323-bfad-47a1-84a5-7b51025f5d0d-etcd-serving-ca\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.360475 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/09d82c69-039a-4e2a-bcd7-dad26f5c8486-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8flns\" (UID: \"09d82c69-039a-4e2a-bcd7-dad26f5c8486\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8flns" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.360500 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg4wd\" (UniqueName: \"kubernetes.io/projected/ad129323-bfad-47a1-84a5-7b51025f5d0d-kube-api-access-jg4wd\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.360525 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/1a194512-24b8-43b7-89f9-db084d65ee0a-machine-approver-tls\") pod \"machine-approver-56656f9798-mt2gv\" (UID: \"1a194512-24b8-43b7-89f9-db084d65ee0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mt2gv" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.360557 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/92275d4e-f157-49a7-b292-2cd46bc23c11-signing-key\") pod \"service-ca-9c57cc56f-l2z5f\" (UID: \"92275d4e-f157-49a7-b292-2cd46bc23c11\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2z5f" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.360590 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-console-oauth-config\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.361374 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e9ffa547-6ab2-4884-8433-d80fff56b441-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5svf7\" (UID: \"e9ffa547-6ab2-4884-8433-d80fff56b441\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5svf7" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.362361 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rb78z"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.363346 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rb78z" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.363544 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9rgf" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.364684 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.365574 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-service-ca\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.366172 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a194512-24b8-43b7-89f9-db084d65ee0a-config\") pod \"machine-approver-56656f9798-mt2gv\" (UID: \"1a194512-24b8-43b7-89f9-db084d65ee0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mt2gv" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.366307 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dea78529-724d-461e-a64c-ed4858ff983c-config\") pod \"authentication-operator-69f744f599-4s8fg\" (UID: \"dea78529-724d-461e-a64c-ed4858ff983c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4s8fg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.366710 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bec2b13f-9bf4-4378-8a9c-56e61397071a-encryption-config\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.367563 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-audit-policies\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.367976 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ad129323-bfad-47a1-84a5-7b51025f5d0d-encryption-config\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.368582 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2ee8da6-43e0-496d-aea1-2c7849305419-config\") pod \"controller-manager-879f6c89f-hh7tq\" (UID: \"b2ee8da6-43e0-496d-aea1-2c7849305419\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.369971 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dea78529-724d-461e-a64c-ed4858ff983c-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4s8fg\" (UID: \"dea78529-724d-461e-a64c-ed4858ff983c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4s8fg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.370681 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d564c\" (UniqueName: \"kubernetes.io/projected/e9ffa547-6ab2-4884-8433-d80fff56b441-kube-api-access-d564c\") pod \"cluster-image-registry-operator-dc59b4c8b-5svf7\" (UID: \"e9ffa547-6ab2-4884-8433-d80fff56b441\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5svf7" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.370740 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ad129323-bfad-47a1-84a5-7b51025f5d0d-etcd-client\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.370771 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b207ee8-d007-408b-8519-b6fa02401fd0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-b6rr5\" (UID: \"6b207ee8-d007-408b-8519-b6fa02401fd0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b6rr5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.370799 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad129323-bfad-47a1-84a5-7b51025f5d0d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.370839 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h79j6\" (UniqueName: \"kubernetes.io/projected/c48e4a1e-a25a-49d8-aa80-bd1e7425ae33-kube-api-access-h79j6\") pod \"openshift-config-operator-7777fb866f-8htrd\" (UID: \"c48e4a1e-a25a-49d8-aa80-bd1e7425ae33\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8htrd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.370866 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.370898 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec125663-8341-427e-917c-758bbeae4251-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7kdcr\" (UID: \"ec125663-8341-427e-917c-758bbeae4251\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7kdcr" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.370926 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f78b5fb4-0a61-4222-928f-2383bac97358-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-r9hlj\" (UID: \"f78b5fb4-0a61-4222-928f-2383bac97358\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-r9hlj" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.370958 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ad129323-bfad-47a1-84a5-7b51025f5d0d-audit\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.370985 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2f691127-6990-46bd-afa7-d6f3f7eaa329-images\") pod \"machine-api-operator-5694c8668f-zsb7g\" (UID: \"2f691127-6990-46bd-afa7-d6f3f7eaa329\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zsb7g" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.371011 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxnht\" (UniqueName: \"kubernetes.io/projected/92275d4e-f157-49a7-b292-2cd46bc23c11-kube-api-access-qxnht\") pod \"service-ca-9c57cc56f-l2z5f\" (UID: \"92275d4e-f157-49a7-b292-2cd46bc23c11\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2z5f" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.371042 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m4lf\" (UniqueName: \"kubernetes.io/projected/4ef05574-4924-4a73-b0ba-0e632c96410a-kube-api-access-7m4lf\") pod \"marketplace-operator-79b997595-sqbkd\" (UID: \"4ef05574-4924-4a73-b0ba-0e632c96410a\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.371053 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9a11b49f-69d2-4d9d-a451-7b69b797da20-trusted-ca\") pod \"console-operator-58897d9998-bxmjt\" (UID: \"9a11b49f-69d2-4d9d-a451-7b69b797da20\") " pod="openshift-console-operator/console-operator-58897d9998-bxmjt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.371062 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/59bf594f-62b6-4a2f-a50d-30b5d8357133-default-certificate\") pod \"router-default-5444994796-q8kdg\" (UID: \"59bf594f-62b6-4a2f-a50d-30b5d8357133\") " pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.371142 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.371179 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a3d0383-95ba-457c-af8c-0eebe774e2ea-config\") pod \"kube-controller-manager-operator-78b949d7b-gmbjm\" (UID: \"1a3d0383-95ba-457c-af8c-0eebe774e2ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gmbjm" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.371222 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2d27f0e8-6b63-4128-b886-7f0018a6a18b-etcd-service-ca\") pod \"etcd-operator-b45778765-cl28q\" (UID: \"2d27f0e8-6b63-4128-b886-7f0018a6a18b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.371248 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkdx2\" (UniqueName: \"kubernetes.io/projected/9a11b49f-69d2-4d9d-a451-7b69b797da20-kube-api-access-kkdx2\") pod \"console-operator-58897d9998-bxmjt\" (UID: \"9a11b49f-69d2-4d9d-a451-7b69b797da20\") " pod="openshift-console-operator/console-operator-58897d9998-bxmjt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.371275 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkv6n\" (UniqueName: \"kubernetes.io/projected/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-kube-api-access-nkv6n\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.371299 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5581f00e-5f39-454d-843d-9dbf6d2cbb0b-trusted-ca\") pod \"ingress-operator-5b745b69d9-s2z8b\" (UID: \"5581f00e-5f39-454d-843d-9dbf6d2cbb0b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.371323 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g28ws\" (UniqueName: \"kubernetes.io/projected/b2ee8da6-43e0-496d-aea1-2c7849305419-kube-api-access-g28ws\") pod \"controller-manager-879f6c89f-hh7tq\" (UID: \"b2ee8da6-43e0-496d-aea1-2c7849305419\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.371348 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59bf594f-62b6-4a2f-a50d-30b5d8357133-metrics-certs\") pod \"router-default-5444994796-q8kdg\" (UID: \"59bf594f-62b6-4a2f-a50d-30b5d8357133\") " pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.371420 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/bc81f2fa-5ee7-4034-81d3-a11639762ac9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-cxghf\" (UID: \"bc81f2fa-5ee7-4034-81d3-a11639762ac9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cxghf" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.371593 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1a3d0383-95ba-457c-af8c-0eebe774e2ea-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gmbjm\" (UID: \"1a3d0383-95ba-457c-af8c-0eebe774e2ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gmbjm" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.371801 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5581f00e-5f39-454d-843d-9dbf6d2cbb0b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-s2z8b\" (UID: \"5581f00e-5f39-454d-843d-9dbf6d2cbb0b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.371869 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bec2b13f-9bf4-4378-8a9c-56e61397071a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.371919 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7f5b\" (UniqueName: \"kubernetes.io/projected/5581f00e-5f39-454d-843d-9dbf6d2cbb0b-kube-api-access-v7f5b\") pod \"ingress-operator-5b745b69d9-s2z8b\" (UID: \"5581f00e-5f39-454d-843d-9dbf6d2cbb0b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.371947 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db43e12f-df81-44cf-9220-2f75d5ecc70b-config\") pod \"kube-apiserver-operator-766d6c64bb-xctz5\" (UID: \"db43e12f-df81-44cf-9220-2f75d5ecc70b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xctz5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.371970 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4ef05574-4924-4a73-b0ba-0e632c96410a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sqbkd\" (UID: \"4ef05574-4924-4a73-b0ba-0e632c96410a\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.372003 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xwtq\" (UniqueName: \"kubernetes.io/projected/959dff27-2ebe-417a-b5be-3e814dff3797-kube-api-access-6xwtq\") pod \"kube-storage-version-migrator-operator-b67b599dd-4qsl8\" (UID: \"959dff27-2ebe-417a-b5be-3e814dff3797\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4qsl8" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.372080 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2d27f0e8-6b63-4128-b886-7f0018a6a18b-etcd-ca\") pod \"etcd-operator-b45778765-cl28q\" (UID: \"2d27f0e8-6b63-4128-b886-7f0018a6a18b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.372165 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.372329 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-console-config\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.372191 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.373326 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.373352 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d27f0e8-6b63-4128-b886-7f0018a6a18b-serving-cert\") pod \"etcd-operator-b45778765-cl28q\" (UID: \"2d27f0e8-6b63-4128-b886-7f0018a6a18b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.373375 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad129323-bfad-47a1-84a5-7b51025f5d0d-serving-cert\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.373399 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-console-serving-cert\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.373421 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec125663-8341-427e-917c-758bbeae4251-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7kdcr\" (UID: \"ec125663-8341-427e-917c-758bbeae4251\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7kdcr" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.373457 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6f3f593c-f784-4f1c-8fb3-016275c38649-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-hpc5d\" (UID: \"6f3f593c-f784-4f1c-8fb3-016275c38649\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hpc5d" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.373484 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/185f4f54-9f50-4017-a7fc-7c7821338442-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2cj4m\" (UID: \"185f4f54-9f50-4017-a7fc-7c7821338442\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2cj4m" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.373509 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ad129323-bfad-47a1-84a5-7b51025f5d0d-image-import-ca\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.373756 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/59bf594f-62b6-4a2f-a50d-30b5d8357133-service-ca-bundle\") pod \"router-default-5444994796-q8kdg\" (UID: \"59bf594f-62b6-4a2f-a50d-30b5d8357133\") " pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.373777 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec125663-8341-427e-917c-758bbeae4251-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7kdcr\" (UID: \"ec125663-8341-427e-917c-758bbeae4251\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7kdcr" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.373779 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dea78529-724d-461e-a64c-ed4858ff983c-serving-cert\") pod \"authentication-operator-69f744f599-4s8fg\" (UID: \"dea78529-724d-461e-a64c-ed4858ff983c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4s8fg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.373801 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/db43e12f-df81-44cf-9220-2f75d5ecc70b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xctz5\" (UID: \"db43e12f-df81-44cf-9220-2f75d5ecc70b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xctz5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.373824 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6c98467a-2f3b-4329-8f21-9492b087b5d8-metrics-tls\") pod \"dns-operator-744455d44c-5bv45\" (UID: \"6c98467a-2f3b-4329-8f21-9492b087b5d8\") " pod="openshift-dns-operator/dns-operator-744455d44c-5bv45" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.373846 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.373866 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2d27f0e8-6b63-4128-b886-7f0018a6a18b-etcd-client\") pod \"etcd-operator-b45778765-cl28q\" (UID: \"2d27f0e8-6b63-4128-b886-7f0018a6a18b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.373885 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vj4m\" (UniqueName: \"kubernetes.io/projected/185f4f54-9f50-4017-a7fc-7c7821338442-kube-api-access-4vj4m\") pod \"openshift-apiserver-operator-796bbdcf4f-2cj4m\" (UID: \"185f4f54-9f50-4017-a7fc-7c7821338442\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2cj4m" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.373905 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5g25\" (UniqueName: \"kubernetes.io/projected/2d27f0e8-6b63-4128-b886-7f0018a6a18b-kube-api-access-n5g25\") pod \"etcd-operator-b45778765-cl28q\" (UID: \"2d27f0e8-6b63-4128-b886-7f0018a6a18b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.373927 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d27f0e8-6b63-4128-b886-7f0018a6a18b-config\") pod \"etcd-operator-b45778765-cl28q\" (UID: \"2d27f0e8-6b63-4128-b886-7f0018a6a18b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.373946 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2ee8da6-43e0-496d-aea1-2c7849305419-serving-cert\") pod \"controller-manager-879f6c89f-hh7tq\" (UID: \"b2ee8da6-43e0-496d-aea1-2c7849305419\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.373991 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ee3b2702-a354-4921-a66d-650bedd0c1b2-profile-collector-cert\") pod \"catalog-operator-68c6474976-l9fbz\" (UID: \"ee3b2702-a354-4921-a66d-650bedd0c1b2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.374014 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5vnm\" (UniqueName: \"kubernetes.io/projected/6c98467a-2f3b-4329-8f21-9492b087b5d8-kube-api-access-n5vnm\") pod \"dns-operator-744455d44c-5bv45\" (UID: \"6c98467a-2f3b-4329-8f21-9492b087b5d8\") " pod="openshift-dns-operator/dns-operator-744455d44c-5bv45" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.374477 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-trusted-ca-bundle\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.374793 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-client-ca\") pod \"route-controller-manager-6576b87f9c-t9q7n\" (UID: \"5a1d0b1c-9bae-483f-886d-bf4acffee7d9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.382566 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.383423 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4s8fg"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.385244 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-audit-dir\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.385310 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ad129323-bfad-47a1-84a5-7b51025f5d0d-node-pullsecrets\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.385342 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ad129323-bfad-47a1-84a5-7b51025f5d0d-audit-dir\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.385801 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bec2b13f-9bf4-4378-8a9c-56e61397071a-audit-dir\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.386777 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.389036 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.389862 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.390768 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.393935 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1a194512-24b8-43b7-89f9-db084d65ee0a-auth-proxy-config\") pod \"machine-approver-56656f9798-mt2gv\" (UID: \"1a194512-24b8-43b7-89f9-db084d65ee0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mt2gv" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.397683 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.397848 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-serving-cert\") pod \"route-controller-manager-6576b87f9c-t9q7n\" (UID: \"5a1d0b1c-9bae-483f-886d-bf4acffee7d9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.398270 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.398290 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.398896 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6c98467a-2f3b-4329-8f21-9492b087b5d8-metrics-tls\") pod \"dns-operator-744455d44c-5bv45\" (UID: \"6c98467a-2f3b-4329-8f21-9492b087b5d8\") " pod="openshift-dns-operator/dns-operator-744455d44c-5bv45" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.399255 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bec2b13f-9bf4-4378-8a9c-56e61397071a-serving-cert\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.400292 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cxghf"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.400961 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cl28q"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.401589 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bec2b13f-9bf4-4378-8a9c-56e61397071a-etcd-client\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.401974 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/2f691127-6990-46bd-afa7-d6f3f7eaa329-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-zsb7g\" (UID: \"2f691127-6990-46bd-afa7-d6f3f7eaa329\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zsb7g" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.402239 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2f691127-6990-46bd-afa7-d6f3f7eaa329-images\") pod \"machine-api-operator-5694c8668f-zsb7g\" (UID: \"2f691127-6990-46bd-afa7-d6f3f7eaa329\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zsb7g" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.402750 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad129323-bfad-47a1-84a5-7b51025f5d0d-serving-cert\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.402810 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ad129323-bfad-47a1-84a5-7b51025f5d0d-etcd-client\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.404158 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-z7c9x"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.405428 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-z7c9x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.405464 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-wmkdd"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.406361 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/bc81f2fa-5ee7-4034-81d3-a11639762ac9-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-cxghf\" (UID: \"bc81f2fa-5ee7-4034-81d3-a11639762ac9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cxghf" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.411147 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bec2b13f-9bf4-4378-8a9c-56e61397071a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.411435 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.411879 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bec2b13f-9bf4-4378-8a9c-56e61397071a-audit-policies\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.412502 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bec2b13f-9bf4-4378-8a9c-56e61397071a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.414656 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d27f0e8-6b63-4128-b886-7f0018a6a18b-serving-cert\") pod \"etcd-operator-b45778765-cl28q\" (UID: \"2d27f0e8-6b63-4128-b886-7f0018a6a18b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.417285 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a11b49f-69d2-4d9d-a451-7b69b797da20-serving-cert\") pod \"console-operator-58897d9998-bxmjt\" (UID: \"9a11b49f-69d2-4d9d-a451-7b69b797da20\") " pod="openshift-console-operator/console-operator-58897d9998-bxmjt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.417496 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b6rr5"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.417913 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.418328 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b207ee8-d007-408b-8519-b6fa02401fd0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-b6rr5\" (UID: \"6b207ee8-d007-408b-8519-b6fa02401fd0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b6rr5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.419736 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/185f4f54-9f50-4017-a7fc-7c7821338442-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2cj4m\" (UID: \"185f4f54-9f50-4017-a7fc-7c7821338442\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2cj4m" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.420574 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-8htrd"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.421034 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad129323-bfad-47a1-84a5-7b51025f5d0d-config\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.421051 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ad129323-bfad-47a1-84a5-7b51025f5d0d-etcd-serving-ca\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.421479 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/185f4f54-9f50-4017-a7fc-7c7821338442-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2cj4m\" (UID: \"185f4f54-9f50-4017-a7fc-7c7821338442\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2cj4m" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.422237 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.422971 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ad129323-bfad-47a1-84a5-7b51025f5d0d-audit\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.423237 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ad129323-bfad-47a1-84a5-7b51025f5d0d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.423443 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d27f0e8-6b63-4128-b886-7f0018a6a18b-config\") pod \"etcd-operator-b45778765-cl28q\" (UID: \"2d27f0e8-6b63-4128-b886-7f0018a6a18b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.423956 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-console-oauth-config\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.423999 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2d27f0e8-6b63-4128-b886-7f0018a6a18b-etcd-ca\") pod \"etcd-operator-b45778765-cl28q\" (UID: \"2d27f0e8-6b63-4128-b886-7f0018a6a18b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.424021 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.424425 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2d27f0e8-6b63-4128-b886-7f0018a6a18b-etcd-service-ca\") pod \"etcd-operator-b45778765-cl28q\" (UID: \"2d27f0e8-6b63-4128-b886-7f0018a6a18b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.424580 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-console-serving-cert\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.425229 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c48e4a1e-a25a-49d8-aa80-bd1e7425ae33-serving-cert\") pod \"openshift-config-operator-7777fb866f-8htrd\" (UID: \"c48e4a1e-a25a-49d8-aa80-bd1e7425ae33\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8htrd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.425382 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ad129323-bfad-47a1-84a5-7b51025f5d0d-image-import-ca\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.427031 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.427566 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/1a194512-24b8-43b7-89f9-db084d65ee0a-machine-approver-tls\") pod \"machine-approver-56656f9798-mt2gv\" (UID: \"1a194512-24b8-43b7-89f9-db084d65ee0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mt2gv" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.428299 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2ee8da6-43e0-496d-aea1-2c7849305419-serving-cert\") pod \"controller-manager-879f6c89f-hh7tq\" (UID: \"b2ee8da6-43e0-496d-aea1-2c7849305419\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.429685 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-bxmjt"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.430498 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-r9hlj"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.431096 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.432567 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.434155 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2d27f0e8-6b63-4128-b886-7f0018a6a18b-etcd-client\") pod \"etcd-operator-b45778765-cl28q\" (UID: \"2d27f0e8-6b63-4128-b886-7f0018a6a18b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.435004 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5xxhk"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.437203 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.438547 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xctz5"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.439844 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-5bv45"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.441743 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4qsl8"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.442995 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-m68m5"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.444429 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-l2z5f"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.445448 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zsb7g"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.447352 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-fgghg"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.448034 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-fgghg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.448498 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-rtlhn"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.449078 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-rtlhn" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.449464 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-d2rzv"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.450852 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gmbjm"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.453225 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-77kj8"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.454665 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.454752 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8flns"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.455777 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-t9rgf"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.456821 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rb78z"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.458393 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-config\") pod \"route-controller-manager-6576b87f9c-t9q7n\" (UID: \"5a1d0b1c-9bae-483f-886d-bf4acffee7d9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.458658 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7kdcr"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.459932 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5svf7"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.461211 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.462193 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sqbkd"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.463331 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hpc5d"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.466596 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.468667 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-z7c9x"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.469079 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-fgghg"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.470299 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.470444 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.472719 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wc8cd"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.474114 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wc8cd"] Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.474279 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.475500 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xx2d\" (UniqueName: \"kubernetes.io/projected/ee3b2702-a354-4921-a66d-650bedd0c1b2-kube-api-access-8xx2d\") pod \"catalog-operator-68c6474976-l9fbz\" (UID: \"ee3b2702-a354-4921-a66d-650bedd0c1b2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.475610 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ee3b2702-a354-4921-a66d-650bedd0c1b2-srv-cert\") pod \"catalog-operator-68c6474976-l9fbz\" (UID: \"ee3b2702-a354-4921-a66d-650bedd0c1b2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.475713 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4ef05574-4924-4a73-b0ba-0e632c96410a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sqbkd\" (UID: \"4ef05574-4924-4a73-b0ba-0e632c96410a\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.475784 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/959dff27-2ebe-417a-b5be-3e814dff3797-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-4qsl8\" (UID: \"959dff27-2ebe-417a-b5be-3e814dff3797\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4qsl8" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.475868 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/09d82c69-039a-4e2a-bcd7-dad26f5c8486-images\") pod \"machine-config-operator-74547568cd-8flns\" (UID: \"09d82c69-039a-4e2a-bcd7-dad26f5c8486\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8flns" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.475938 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db43e12f-df81-44cf-9220-2f75d5ecc70b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xctz5\" (UID: \"db43e12f-df81-44cf-9220-2f75d5ecc70b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xctz5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.476020 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/92275d4e-f157-49a7-b292-2cd46bc23c11-signing-cabundle\") pod \"service-ca-9c57cc56f-l2z5f\" (UID: \"92275d4e-f157-49a7-b292-2cd46bc23c11\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2z5f" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.476098 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7wjj\" (UniqueName: \"kubernetes.io/projected/6f3f593c-f784-4f1c-8fb3-016275c38649-kube-api-access-n7wjj\") pod \"control-plane-machine-set-operator-78cbb6b69f-hpc5d\" (UID: \"6f3f593c-f784-4f1c-8fb3-016275c38649\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hpc5d" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.476167 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5581f00e-5f39-454d-843d-9dbf6d2cbb0b-metrics-tls\") pod \"ingress-operator-5b745b69d9-s2z8b\" (UID: \"5581f00e-5f39-454d-843d-9dbf6d2cbb0b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.476235 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c5bd\" (UniqueName: \"kubernetes.io/projected/f78b5fb4-0a61-4222-928f-2383bac97358-kube-api-access-7c5bd\") pod \"multus-admission-controller-857f4d67dd-r9hlj\" (UID: \"f78b5fb4-0a61-4222-928f-2383bac97358\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-r9hlj" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.476308 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/59bf594f-62b6-4a2f-a50d-30b5d8357133-stats-auth\") pod \"router-default-5444994796-q8kdg\" (UID: \"59bf594f-62b6-4a2f-a50d-30b5d8357133\") " pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.476404 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjmr7\" (UniqueName: \"kubernetes.io/projected/09d82c69-039a-4e2a-bcd7-dad26f5c8486-kube-api-access-kjmr7\") pod \"machine-config-operator-74547568cd-8flns\" (UID: \"09d82c69-039a-4e2a-bcd7-dad26f5c8486\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8flns" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.476497 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/09d82c69-039a-4e2a-bcd7-dad26f5c8486-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8flns\" (UID: \"09d82c69-039a-4e2a-bcd7-dad26f5c8486\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8flns" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.476609 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/92275d4e-f157-49a7-b292-2cd46bc23c11-signing-key\") pod \"service-ca-9c57cc56f-l2z5f\" (UID: \"92275d4e-f157-49a7-b292-2cd46bc23c11\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2z5f" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.476712 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec125663-8341-427e-917c-758bbeae4251-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7kdcr\" (UID: \"ec125663-8341-427e-917c-758bbeae4251\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7kdcr" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.476798 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f78b5fb4-0a61-4222-928f-2383bac97358-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-r9hlj\" (UID: \"f78b5fb4-0a61-4222-928f-2383bac97358\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-r9hlj" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.476914 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxnht\" (UniqueName: \"kubernetes.io/projected/92275d4e-f157-49a7-b292-2cd46bc23c11-kube-api-access-qxnht\") pod \"service-ca-9c57cc56f-l2z5f\" (UID: \"92275d4e-f157-49a7-b292-2cd46bc23c11\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2z5f" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.477006 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m4lf\" (UniqueName: \"kubernetes.io/projected/4ef05574-4924-4a73-b0ba-0e632c96410a-kube-api-access-7m4lf\") pod \"marketplace-operator-79b997595-sqbkd\" (UID: \"4ef05574-4924-4a73-b0ba-0e632c96410a\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.477112 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a3d0383-95ba-457c-af8c-0eebe774e2ea-config\") pod \"kube-controller-manager-operator-78b949d7b-gmbjm\" (UID: \"1a3d0383-95ba-457c-af8c-0eebe774e2ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gmbjm" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.477190 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/59bf594f-62b6-4a2f-a50d-30b5d8357133-default-certificate\") pod \"router-default-5444994796-q8kdg\" (UID: \"59bf594f-62b6-4a2f-a50d-30b5d8357133\") " pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.477268 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5581f00e-5f39-454d-843d-9dbf6d2cbb0b-trusted-ca\") pod \"ingress-operator-5b745b69d9-s2z8b\" (UID: \"5581f00e-5f39-454d-843d-9dbf6d2cbb0b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.477358 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/09d82c69-039a-4e2a-bcd7-dad26f5c8486-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8flns\" (UID: \"09d82c69-039a-4e2a-bcd7-dad26f5c8486\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8flns" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.477458 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59bf594f-62b6-4a2f-a50d-30b5d8357133-metrics-certs\") pod \"router-default-5444994796-q8kdg\" (UID: \"59bf594f-62b6-4a2f-a50d-30b5d8357133\") " pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.477558 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1a3d0383-95ba-457c-af8c-0eebe774e2ea-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gmbjm\" (UID: \"1a3d0383-95ba-457c-af8c-0eebe774e2ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gmbjm" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.477649 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5581f00e-5f39-454d-843d-9dbf6d2cbb0b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-s2z8b\" (UID: \"5581f00e-5f39-454d-843d-9dbf6d2cbb0b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.477720 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7f5b\" (UniqueName: \"kubernetes.io/projected/5581f00e-5f39-454d-843d-9dbf6d2cbb0b-kube-api-access-v7f5b\") pod \"ingress-operator-5b745b69d9-s2z8b\" (UID: \"5581f00e-5f39-454d-843d-9dbf6d2cbb0b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.477812 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db43e12f-df81-44cf-9220-2f75d5ecc70b-config\") pod \"kube-apiserver-operator-766d6c64bb-xctz5\" (UID: \"db43e12f-df81-44cf-9220-2f75d5ecc70b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xctz5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.477902 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4ef05574-4924-4a73-b0ba-0e632c96410a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sqbkd\" (UID: \"4ef05574-4924-4a73-b0ba-0e632c96410a\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.478043 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xwtq\" (UniqueName: \"kubernetes.io/projected/959dff27-2ebe-417a-b5be-3e814dff3797-kube-api-access-6xwtq\") pod \"kube-storage-version-migrator-operator-b67b599dd-4qsl8\" (UID: \"959dff27-2ebe-417a-b5be-3e814dff3797\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4qsl8" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.478122 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec125663-8341-427e-917c-758bbeae4251-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7kdcr\" (UID: \"ec125663-8341-427e-917c-758bbeae4251\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7kdcr" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.478234 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6f3f593c-f784-4f1c-8fb3-016275c38649-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-hpc5d\" (UID: \"6f3f593c-f784-4f1c-8fb3-016275c38649\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hpc5d" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.478320 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/59bf594f-62b6-4a2f-a50d-30b5d8357133-service-ca-bundle\") pod \"router-default-5444994796-q8kdg\" (UID: \"59bf594f-62b6-4a2f-a50d-30b5d8357133\") " pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.478387 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec125663-8341-427e-917c-758bbeae4251-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7kdcr\" (UID: \"ec125663-8341-427e-917c-758bbeae4251\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7kdcr" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.478470 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/db43e12f-df81-44cf-9220-2f75d5ecc70b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xctz5\" (UID: \"db43e12f-df81-44cf-9220-2f75d5ecc70b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xctz5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.478595 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ee3b2702-a354-4921-a66d-650bedd0c1b2-profile-collector-cert\") pod \"catalog-operator-68c6474976-l9fbz\" (UID: \"ee3b2702-a354-4921-a66d-650bedd0c1b2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.478685 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-m68m5\" (UID: \"fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m68m5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.478786 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjddt\" (UniqueName: \"kubernetes.io/projected/59bf594f-62b6-4a2f-a50d-30b5d8357133-kube-api-access-fjddt\") pod \"router-default-5444994796-q8kdg\" (UID: \"59bf594f-62b6-4a2f-a50d-30b5d8357133\") " pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.478880 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a3d0383-95ba-457c-af8c-0eebe774e2ea-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gmbjm\" (UID: \"1a3d0383-95ba-457c-af8c-0eebe774e2ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gmbjm" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.478985 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/09d82c69-039a-4e2a-bcd7-dad26f5c8486-proxy-tls\") pod \"machine-config-operator-74547568cd-8flns\" (UID: \"09d82c69-039a-4e2a-bcd7-dad26f5c8486\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8flns" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.479062 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/959dff27-2ebe-417a-b5be-3e814dff3797-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-4qsl8\" (UID: \"959dff27-2ebe-417a-b5be-3e814dff3797\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4qsl8" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.479130 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc-proxy-tls\") pod \"machine-config-controller-84d6567774-m68m5\" (UID: \"fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m68m5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.479194 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec125663-8341-427e-917c-758bbeae4251-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7kdcr\" (UID: \"ec125663-8341-427e-917c-758bbeae4251\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7kdcr" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.479263 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdtwh\" (UniqueName: \"kubernetes.io/projected/fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc-kube-api-access-rdtwh\") pod \"machine-config-controller-84d6567774-m68m5\" (UID: \"fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m68m5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.479364 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7wfh\" (UniqueName: \"kubernetes.io/projected/7d9d9160-dd25-4640-a84b-ee9558be3b88-kube-api-access-j7wfh\") pod \"migrator-59844c95c7-77kj8\" (UID: \"7d9d9160-dd25-4640-a84b-ee9558be3b88\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-77kj8" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.479478 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-m68m5\" (UID: \"fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m68m5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.490391 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.510691 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.522229 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec125663-8341-427e-917c-758bbeae4251-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7kdcr\" (UID: \"ec125663-8341-427e-917c-758bbeae4251\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7kdcr" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.531126 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.551374 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.565742 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/959dff27-2ebe-417a-b5be-3e814dff3797-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-4qsl8\" (UID: \"959dff27-2ebe-417a-b5be-3e814dff3797\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4qsl8" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.571860 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.590376 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.610585 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.627892 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db43e12f-df81-44cf-9220-2f75d5ecc70b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xctz5\" (UID: \"db43e12f-df81-44cf-9220-2f75d5ecc70b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xctz5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.631737 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.639452 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db43e12f-df81-44cf-9220-2f75d5ecc70b-config\") pod \"kube-apiserver-operator-766d6c64bb-xctz5\" (UID: \"db43e12f-df81-44cf-9220-2f75d5ecc70b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xctz5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.646142 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"1b4f6835b8c873443445bf3eb88cf56111a291e8fc92488a0853796acacbb41f"} Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.649807 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"8e7a0c448adc6a6811d429a1239679c7bc00b5b50303ebd461763113620f1497"} Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.649831 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7362ab0485cadbf8ce0bee70b920b07fcc929a3aee97eddfbd5553379158b456"} Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.655223 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: W1003 12:53:40.665831 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-c5280b2410740bf7e56b19a5898127fab56a48b32fec6050d5ac381fdfbb6d45 WatchSource:0}: Error finding container c5280b2410740bf7e56b19a5898127fab56a48b32fec6050d5ac381fdfbb6d45: Status 404 returned error can't find the container with id c5280b2410740bf7e56b19a5898127fab56a48b32fec6050d5ac381fdfbb6d45 Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.670423 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.690312 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.696778 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/959dff27-2ebe-417a-b5be-3e814dff3797-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-4qsl8\" (UID: \"959dff27-2ebe-417a-b5be-3e814dff3797\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4qsl8" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.712481 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.730942 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.743873 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/09d82c69-039a-4e2a-bcd7-dad26f5c8486-proxy-tls\") pod \"machine-config-operator-74547568cd-8flns\" (UID: \"09d82c69-039a-4e2a-bcd7-dad26f5c8486\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8flns" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.750463 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.770337 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.790255 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.799221 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5581f00e-5f39-454d-843d-9dbf6d2cbb0b-metrics-tls\") pod \"ingress-operator-5b745b69d9-s2z8b\" (UID: \"5581f00e-5f39-454d-843d-9dbf6d2cbb0b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.814420 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.819921 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5581f00e-5f39-454d-843d-9dbf6d2cbb0b-trusted-ca\") pod \"ingress-operator-5b745b69d9-s2z8b\" (UID: \"5581f00e-5f39-454d-843d-9dbf6d2cbb0b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.830560 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.849981 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.870451 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.877430 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/09d82c69-039a-4e2a-bcd7-dad26f5c8486-images\") pod \"machine-config-operator-74547568cd-8flns\" (UID: \"09d82c69-039a-4e2a-bcd7-dad26f5c8486\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8flns" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.890432 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.902484 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6f3f593c-f784-4f1c-8fb3-016275c38649-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-hpc5d\" (UID: \"6f3f593c-f784-4f1c-8fb3-016275c38649\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hpc5d" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.912012 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.930976 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.954696 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.963220 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc-proxy-tls\") pod \"machine-config-controller-84d6567774-m68m5\" (UID: \"fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m68m5" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.971839 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 03 12:53:40 crc kubenswrapper[4578]: I1003 12:53:40.990213 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.011168 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.030549 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.050911 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.062335 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a3d0383-95ba-457c-af8c-0eebe774e2ea-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gmbjm\" (UID: \"1a3d0383-95ba-457c-af8c-0eebe774e2ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gmbjm" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.070459 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.090449 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.098235 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a3d0383-95ba-457c-af8c-0eebe774e2ea-config\") pod \"kube-controller-manager-operator-78b949d7b-gmbjm\" (UID: \"1a3d0383-95ba-457c-af8c-0eebe774e2ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gmbjm" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.131097 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.141281 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f78b5fb4-0a61-4222-928f-2383bac97358-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-r9hlj\" (UID: \"f78b5fb4-0a61-4222-928f-2383bac97358\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-r9hlj" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.151364 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.170153 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.190995 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.210417 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.219623 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4ef05574-4924-4a73-b0ba-0e632c96410a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sqbkd\" (UID: \"4ef05574-4924-4a73-b0ba-0e632c96410a\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.247201 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.249283 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4ef05574-4924-4a73-b0ba-0e632c96410a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sqbkd\" (UID: \"4ef05574-4924-4a73-b0ba-0e632c96410a\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.250397 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.269278 4578 request.go:700] Waited for 1.003795892s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.270753 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.291205 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.301020 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/59bf594f-62b6-4a2f-a50d-30b5d8357133-default-certificate\") pod \"router-default-5444994796-q8kdg\" (UID: \"59bf594f-62b6-4a2f-a50d-30b5d8357133\") " pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.310431 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.320319 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59bf594f-62b6-4a2f-a50d-30b5d8357133-metrics-certs\") pod \"router-default-5444994796-q8kdg\" (UID: \"59bf594f-62b6-4a2f-a50d-30b5d8357133\") " pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.330417 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.350495 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.359469 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/59bf594f-62b6-4a2f-a50d-30b5d8357133-stats-auth\") pod \"router-default-5444994796-q8kdg\" (UID: \"59bf594f-62b6-4a2f-a50d-30b5d8357133\") " pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.370080 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.390726 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.400267 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/59bf594f-62b6-4a2f-a50d-30b5d8357133-service-ca-bundle\") pod \"router-default-5444994796-q8kdg\" (UID: \"59bf594f-62b6-4a2f-a50d-30b5d8357133\") " pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.412008 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.430525 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.451364 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.470518 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 03 12:53:41 crc kubenswrapper[4578]: E1003 12:53:41.476700 4578 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Oct 03 12:53:41 crc kubenswrapper[4578]: E1003 12:53:41.476718 4578 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 03 12:53:41 crc kubenswrapper[4578]: E1003 12:53:41.476771 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/92275d4e-f157-49a7-b292-2cd46bc23c11-signing-cabundle podName:92275d4e-f157-49a7-b292-2cd46bc23c11 nodeName:}" failed. No retries permitted until 2025-10-03 12:53:41.976749292 +0000 UTC m=+157.775221476 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/92275d4e-f157-49a7-b292-2cd46bc23c11-signing-cabundle") pod "service-ca-9c57cc56f-l2z5f" (UID: "92275d4e-f157-49a7-b292-2cd46bc23c11") : failed to sync configmap cache: timed out waiting for the condition Oct 03 12:53:41 crc kubenswrapper[4578]: E1003 12:53:41.476795 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee3b2702-a354-4921-a66d-650bedd0c1b2-srv-cert podName:ee3b2702-a354-4921-a66d-650bedd0c1b2 nodeName:}" failed. No retries permitted until 2025-10-03 12:53:41.976785254 +0000 UTC m=+157.775257438 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/ee3b2702-a354-4921-a66d-650bedd0c1b2-srv-cert") pod "catalog-operator-68c6474976-l9fbz" (UID: "ee3b2702-a354-4921-a66d-650bedd0c1b2") : failed to sync secret cache: timed out waiting for the condition Oct 03 12:53:41 crc kubenswrapper[4578]: E1003 12:53:41.477813 4578 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Oct 03 12:53:41 crc kubenswrapper[4578]: E1003 12:53:41.477934 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/92275d4e-f157-49a7-b292-2cd46bc23c11-signing-key podName:92275d4e-f157-49a7-b292-2cd46bc23c11 nodeName:}" failed. No retries permitted until 2025-10-03 12:53:41.977877829 +0000 UTC m=+157.776350013 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/92275d4e-f157-49a7-b292-2cd46bc23c11-signing-key") pod "service-ca-9c57cc56f-l2z5f" (UID: "92275d4e-f157-49a7-b292-2cd46bc23c11") : failed to sync secret cache: timed out waiting for the condition Oct 03 12:53:41 crc kubenswrapper[4578]: E1003 12:53:41.478904 4578 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Oct 03 12:53:41 crc kubenswrapper[4578]: E1003 12:53:41.478953 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee3b2702-a354-4921-a66d-650bedd0c1b2-profile-collector-cert podName:ee3b2702-a354-4921-a66d-650bedd0c1b2 nodeName:}" failed. No retries permitted until 2025-10-03 12:53:41.978939043 +0000 UTC m=+157.777411227 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/ee3b2702-a354-4921-a66d-650bedd0c1b2-profile-collector-cert") pod "catalog-operator-68c6474976-l9fbz" (UID: "ee3b2702-a354-4921-a66d-650bedd0c1b2") : failed to sync secret cache: timed out waiting for the condition Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.490122 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.511060 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.530362 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.550986 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.570114 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.591132 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.610555 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.630893 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.652328 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.654026 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"11293dddc7664cb037220c9f4e28255bf650c6d09186dfaad43a85aede5d957f"} Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.656170 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"cb771f620da9e4c1f661b884d25a22f0836e82305dc1eafb8956af35ff7f02cd"} Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.656216 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c5280b2410740bf7e56b19a5898127fab56a48b32fec6050d5ac381fdfbb6d45"} Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.656374 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.691556 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.713048 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.747892 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9ffa547-6ab2-4884-8433-d80fff56b441-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5svf7\" (UID: \"e9ffa547-6ab2-4884-8433-d80fff56b441\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5svf7" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.751224 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.771118 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.803892 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbjhc\" (UniqueName: \"kubernetes.io/projected/2f691127-6990-46bd-afa7-d6f3f7eaa329-kube-api-access-kbjhc\") pod \"machine-api-operator-5694c8668f-zsb7g\" (UID: \"2f691127-6990-46bd-afa7-d6f3f7eaa329\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-zsb7g" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.811056 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.831139 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.850377 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.872240 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.891290 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.910754 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.945052 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9mzs\" (UniqueName: \"kubernetes.io/projected/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-kube-api-access-h9mzs\") pod \"route-controller-manager-6576b87f9c-t9q7n\" (UID: \"5a1d0b1c-9bae-483f-886d-bf4acffee7d9\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.965239 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbxfv\" (UniqueName: \"kubernetes.io/projected/bc81f2fa-5ee7-4034-81d3-a11639762ac9-kube-api-access-vbxfv\") pod \"cluster-samples-operator-665b6dd947-cxghf\" (UID: \"bc81f2fa-5ee7-4034-81d3-a11639762ac9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cxghf" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.966824 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-zsb7g" Oct 03 12:53:41 crc kubenswrapper[4578]: I1003 12:53:41.985706 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjg95\" (UniqueName: \"kubernetes.io/projected/1a194512-24b8-43b7-89f9-db084d65ee0a-kube-api-access-wjg95\") pod \"machine-approver-56656f9798-mt2gv\" (UID: \"1a194512-24b8-43b7-89f9-db084d65ee0a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mt2gv" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.003096 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-468cd\" (UniqueName: \"kubernetes.io/projected/bec2b13f-9bf4-4378-8a9c-56e61397071a-kube-api-access-468cd\") pod \"apiserver-7bbb656c7d-rvkjq\" (UID: \"bec2b13f-9bf4-4378-8a9c-56e61397071a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.003761 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ee3b2702-a354-4921-a66d-650bedd0c1b2-profile-collector-cert\") pod \"catalog-operator-68c6474976-l9fbz\" (UID: \"ee3b2702-a354-4921-a66d-650bedd0c1b2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.003864 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ee3b2702-a354-4921-a66d-650bedd0c1b2-srv-cert\") pod \"catalog-operator-68c6474976-l9fbz\" (UID: \"ee3b2702-a354-4921-a66d-650bedd0c1b2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.003923 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/92275d4e-f157-49a7-b292-2cd46bc23c11-signing-cabundle\") pod \"service-ca-9c57cc56f-l2z5f\" (UID: \"92275d4e-f157-49a7-b292-2cd46bc23c11\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2z5f" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.003983 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/92275d4e-f157-49a7-b292-2cd46bc23c11-signing-key\") pod \"service-ca-9c57cc56f-l2z5f\" (UID: \"92275d4e-f157-49a7-b292-2cd46bc23c11\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2z5f" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.005012 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/92275d4e-f157-49a7-b292-2cd46bc23c11-signing-cabundle\") pod \"service-ca-9c57cc56f-l2z5f\" (UID: \"92275d4e-f157-49a7-b292-2cd46bc23c11\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2z5f" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.007099 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ee3b2702-a354-4921-a66d-650bedd0c1b2-profile-collector-cert\") pod \"catalog-operator-68c6474976-l9fbz\" (UID: \"ee3b2702-a354-4921-a66d-650bedd0c1b2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.007878 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ee3b2702-a354-4921-a66d-650bedd0c1b2-srv-cert\") pod \"catalog-operator-68c6474976-l9fbz\" (UID: \"ee3b2702-a354-4921-a66d-650bedd0c1b2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.008914 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/92275d4e-f157-49a7-b292-2cd46bc23c11-signing-key\") pod \"service-ca-9c57cc56f-l2z5f\" (UID: \"92275d4e-f157-49a7-b292-2cd46bc23c11\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2z5f" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.020482 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.025504 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cnqw\" (UniqueName: \"kubernetes.io/projected/de78c746-e145-48ba-8601-a3ff2c6b2b62-kube-api-access-9cnqw\") pod \"downloads-7954f5f757-d2rzv\" (UID: \"de78c746-e145-48ba-8601-a3ff2c6b2b62\") " pod="openshift-console/downloads-7954f5f757-d2rzv" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.048695 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6j4z\" (UniqueName: \"kubernetes.io/projected/dea78529-724d-461e-a64c-ed4858ff983c-kube-api-access-m6j4z\") pod \"authentication-operator-69f744f599-4s8fg\" (UID: \"dea78529-724d-461e-a64c-ed4858ff983c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4s8fg" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.064616 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5vnm\" (UniqueName: \"kubernetes.io/projected/6c98467a-2f3b-4329-8f21-9492b087b5d8-kube-api-access-n5vnm\") pod \"dns-operator-744455d44c-5bv45\" (UID: \"6c98467a-2f3b-4329-8f21-9492b087b5d8\") " pod="openshift-dns-operator/dns-operator-744455d44c-5bv45" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.079029 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cxghf" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.101401 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjfxr\" (UniqueName: \"kubernetes.io/projected/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-kube-api-access-bjfxr\") pod \"console-f9d7485db-ns26l\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.107223 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mt2gv" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.114314 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkdx2\" (UniqueName: \"kubernetes.io/projected/9a11b49f-69d2-4d9d-a451-7b69b797da20-kube-api-access-kkdx2\") pod \"console-operator-58897d9998-bxmjt\" (UID: \"9a11b49f-69d2-4d9d-a451-7b69b797da20\") " pod="openshift-console-operator/console-operator-58897d9998-bxmjt" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.130735 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkv6n\" (UniqueName: \"kubernetes.io/projected/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-kube-api-access-nkv6n\") pod \"oauth-openshift-558db77b4-7br6x\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.154321 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g28ws\" (UniqueName: \"kubernetes.io/projected/b2ee8da6-43e0-496d-aea1-2c7849305419-kube-api-access-g28ws\") pod \"controller-manager-879f6c89f-hh7tq\" (UID: \"b2ee8da6-43e0-496d-aea1-2c7849305419\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.176184 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg4wd\" (UniqueName: \"kubernetes.io/projected/ad129323-bfad-47a1-84a5-7b51025f5d0d-kube-api-access-jg4wd\") pod \"apiserver-76f77b778f-wmkdd\" (UID: \"ad129323-bfad-47a1-84a5-7b51025f5d0d\") " pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.205626 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vj4m\" (UniqueName: \"kubernetes.io/projected/185f4f54-9f50-4017-a7fc-7c7821338442-kube-api-access-4vj4m\") pod \"openshift-apiserver-operator-796bbdcf4f-2cj4m\" (UID: \"185f4f54-9f50-4017-a7fc-7c7821338442\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2cj4m" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.215392 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-4s8fg" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.225816 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.234152 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5g25\" (UniqueName: \"kubernetes.io/projected/2d27f0e8-6b63-4128-b886-7f0018a6a18b-kube-api-access-n5g25\") pod \"etcd-operator-b45778765-cl28q\" (UID: \"2d27f0e8-6b63-4128-b886-7f0018a6a18b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.243212 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.246568 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpc5w\" (UniqueName: \"kubernetes.io/projected/6b207ee8-d007-408b-8519-b6fa02401fd0-kube-api-access-jpc5w\") pod \"openshift-controller-manager-operator-756b6f6bc6-b6rr5\" (UID: \"6b207ee8-d007-408b-8519-b6fa02401fd0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b6rr5" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.264276 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d564c\" (UniqueName: \"kubernetes.io/projected/e9ffa547-6ab2-4884-8433-d80fff56b441-kube-api-access-d564c\") pod \"cluster-image-registry-operator-dc59b4c8b-5svf7\" (UID: \"e9ffa547-6ab2-4884-8433-d80fff56b441\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5svf7" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.264914 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.267872 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-zsb7g"] Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.269671 4578 request.go:700] Waited for 1.863931892s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/configmaps?fieldSelector=metadata.name%3Ddns-default&limit=500&resourceVersion=0 Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.270823 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.273233 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b6rr5" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.280878 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h79j6\" (UniqueName: \"kubernetes.io/projected/c48e4a1e-a25a-49d8-aa80-bd1e7425ae33-kube-api-access-h79j6\") pod \"openshift-config-operator-7777fb866f-8htrd\" (UID: \"c48e4a1e-a25a-49d8-aa80-bd1e7425ae33\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8htrd" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.283076 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-d2rzv" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.286743 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2cj4m" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.294245 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-bxmjt" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.295000 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.303269 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5svf7" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.306353 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.309847 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.313916 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.313981 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-5bv45" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.332613 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.361006 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.372505 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.376709 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cxghf"] Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.390321 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.391885 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.394691 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n"] Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.412582 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.432167 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.455055 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.479583 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.481305 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4s8fg"] Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.492262 4578 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.511107 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.555053 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8htrd" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.569403 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xx2d\" (UniqueName: \"kubernetes.io/projected/ee3b2702-a354-4921-a66d-650bedd0c1b2-kube-api-access-8xx2d\") pod \"catalog-operator-68c6474976-l9fbz\" (UID: \"ee3b2702-a354-4921-a66d-650bedd0c1b2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.588441 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7wjj\" (UniqueName: \"kubernetes.io/projected/6f3f593c-f784-4f1c-8fb3-016275c38649-kube-api-access-n7wjj\") pod \"control-plane-machine-set-operator-78cbb6b69f-hpc5d\" (UID: \"6f3f593c-f784-4f1c-8fb3-016275c38649\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hpc5d" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.609731 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c5bd\" (UniqueName: \"kubernetes.io/projected/f78b5fb4-0a61-4222-928f-2383bac97358-kube-api-access-7c5bd\") pod \"multus-admission-controller-857f4d67dd-r9hlj\" (UID: \"f78b5fb4-0a61-4222-928f-2383bac97358\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-r9hlj" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.618588 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjmr7\" (UniqueName: \"kubernetes.io/projected/09d82c69-039a-4e2a-bcd7-dad26f5c8486-kube-api-access-kjmr7\") pod \"machine-config-operator-74547568cd-8flns\" (UID: \"09d82c69-039a-4e2a-bcd7-dad26f5c8486\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8flns" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.643415 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxnht\" (UniqueName: \"kubernetes.io/projected/92275d4e-f157-49a7-b292-2cd46bc23c11-kube-api-access-qxnht\") pod \"service-ca-9c57cc56f-l2z5f\" (UID: \"92275d4e-f157-49a7-b292-2cd46bc23c11\") " pod="openshift-service-ca/service-ca-9c57cc56f-l2z5f" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.654814 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8flns" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.658405 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec125663-8341-427e-917c-758bbeae4251-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7kdcr\" (UID: \"ec125663-8341-427e-917c-758bbeae4251\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7kdcr" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.666484 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m4lf\" (UniqueName: \"kubernetes.io/projected/4ef05574-4924-4a73-b0ba-0e632c96410a-kube-api-access-7m4lf\") pod \"marketplace-operator-79b997595-sqbkd\" (UID: \"4ef05574-4924-4a73-b0ba-0e632c96410a\") " pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.681538 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hpc5d" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.689376 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1a3d0383-95ba-457c-af8c-0eebe774e2ea-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gmbjm\" (UID: \"1a3d0383-95ba-457c-af8c-0eebe774e2ea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gmbjm" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.707685 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5581f00e-5f39-454d-843d-9dbf6d2cbb0b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-s2z8b\" (UID: \"5581f00e-5f39-454d-843d-9dbf6d2cbb0b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.711217 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zsb7g" event={"ID":"2f691127-6990-46bd-afa7-d6f3f7eaa329","Type":"ContainerStarted","Data":"742fb4681ddaac3f63c5e1ae9f5f9b518dc58ed95e96b9357e3bacb5988eec55"} Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.716808 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gmbjm" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.724254 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-r9hlj" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.733360 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mt2gv" event={"ID":"1a194512-24b8-43b7-89f9-db084d65ee0a","Type":"ContainerStarted","Data":"d7fab94fab807d698f9b84eca3ff36b75dd82a9d87025856ce982c6eddb8976f"} Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.733402 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mt2gv" event={"ID":"1a194512-24b8-43b7-89f9-db084d65ee0a","Type":"ContainerStarted","Data":"dc2e0e6ed8548b4045d9d0404395fa54269bc3a8558bc1592f34fb75861dbf8c"} Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.735914 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.738644 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7f5b\" (UniqueName: \"kubernetes.io/projected/5581f00e-5f39-454d-843d-9dbf6d2cbb0b-kube-api-access-v7f5b\") pod \"ingress-operator-5b745b69d9-s2z8b\" (UID: \"5581f00e-5f39-454d-843d-9dbf6d2cbb0b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.754024 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xwtq\" (UniqueName: \"kubernetes.io/projected/959dff27-2ebe-417a-b5be-3e814dff3797-kube-api-access-6xwtq\") pod \"kube-storage-version-migrator-operator-b67b599dd-4qsl8\" (UID: \"959dff27-2ebe-417a-b5be-3e814dff3797\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4qsl8" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.758674 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" event={"ID":"5a1d0b1c-9bae-483f-886d-bf4acffee7d9","Type":"ContainerStarted","Data":"e691b4315d1f3f216f70d30ef2ea86c360874b0609153fb8a175555365b37074"} Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.760181 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cxghf" event={"ID":"bc81f2fa-5ee7-4034-81d3-a11639762ac9","Type":"ContainerStarted","Data":"cc8335d3080d5a4f2464803a89c97850a0f41f82e6138c7026be4d5ddce3f2db"} Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.761431 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-4s8fg" event={"ID":"dea78529-724d-461e-a64c-ed4858ff983c","Type":"ContainerStarted","Data":"81aa2b7f4c9406a2b61d91bb6cc5ad0c229013c568a646aaa4fa8997d2b8541c"} Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.764433 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.769204 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/db43e12f-df81-44cf-9220-2f75d5ecc70b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xctz5\" (UID: \"db43e12f-df81-44cf-9220-2f75d5ecc70b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xctz5" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.780787 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-l2z5f" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.782266 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjddt\" (UniqueName: \"kubernetes.io/projected/59bf594f-62b6-4a2f-a50d-30b5d8357133-kube-api-access-fjddt\") pod \"router-default-5444994796-q8kdg\" (UID: \"59bf594f-62b6-4a2f-a50d-30b5d8357133\") " pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.812370 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdtwh\" (UniqueName: \"kubernetes.io/projected/fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc-kube-api-access-rdtwh\") pod \"machine-config-controller-84d6567774-m68m5\" (UID: \"fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m68m5" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.834835 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7wfh\" (UniqueName: \"kubernetes.io/projected/7d9d9160-dd25-4640-a84b-ee9558be3b88-kube-api-access-j7wfh\") pod \"migrator-59844c95c7-77kj8\" (UID: \"7d9d9160-dd25-4640-a84b-ee9558be3b88\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-77kj8" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.904926 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hh7tq"] Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.918317 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-registry-tls\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.918427 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-bound-sa-token\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.918469 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.918729 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.918764 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.918811 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8h85\" (UniqueName: \"kubernetes.io/projected/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-kube-api-access-f8h85\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.918879 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-registry-certificates\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.918953 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-trusted-ca\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:42 crc kubenswrapper[4578]: E1003 12:53:42.920911 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:43.420896306 +0000 UTC m=+159.219368490 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.927620 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7kdcr" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.930387 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq"] Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.942984 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4qsl8" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.943513 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xctz5" Oct 03 12:53:42 crc kubenswrapper[4578]: I1003 12:53:42.972916 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:42.999026 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ns26l"] Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.006778 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m68m5" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.021063 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.021267 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cm8v\" (UniqueName: \"kubernetes.io/projected/ac7d947f-4847-4809-811d-5fc94e141dd4-kube-api-access-5cm8v\") pod \"packageserver-d55dfcdfc-g5m4p\" (UID: \"ac7d947f-4847-4809-811d-5fc94e141dd4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.021331 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4cdt\" (UniqueName: \"kubernetes.io/projected/fdfd4313-d05e-4cbd-8520-3f929675c662-kube-api-access-x4cdt\") pod \"collect-profiles-29324925-64fg4\" (UID: \"fdfd4313-d05e-4cbd-8520-3f929675c662\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.021374 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj2bp\" (UniqueName: \"kubernetes.io/projected/6528b192-d6c1-46e6-8048-58a76459651a-kube-api-access-gj2bp\") pod \"machine-config-server-rtlhn\" (UID: \"6528b192-d6c1-46e6-8048-58a76459651a\") " pod="openshift-machine-config-operator/machine-config-server-rtlhn" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.021416 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ed3f40b1-ee95-427b-a6ea-0b412dba932c-registration-dir\") pod \"csi-hostpathplugin-wc8cd\" (UID: \"ed3f40b1-ee95-427b-a6ea-0b412dba932c\") " pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.021432 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f9658c9-2fb3-44c1-bb34-62a2e16360bb-config\") pod \"service-ca-operator-777779d784-t9rgf\" (UID: \"2f9658c9-2fb3-44c1-bb34-62a2e16360bb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9rgf" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.021467 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6528b192-d6c1-46e6-8048-58a76459651a-certs\") pod \"machine-config-server-rtlhn\" (UID: \"6528b192-d6c1-46e6-8048-58a76459651a\") " pod="openshift-machine-config-operator/machine-config-server-rtlhn" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.021530 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ac7d947f-4847-4809-811d-5fc94e141dd4-tmpfs\") pod \"packageserver-d55dfcdfc-g5m4p\" (UID: \"ac7d947f-4847-4809-811d-5fc94e141dd4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.021574 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-bound-sa-token\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.021590 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbzl5\" (UniqueName: \"kubernetes.io/projected/2f9658c9-2fb3-44c1-bb34-62a2e16360bb-kube-api-access-pbzl5\") pod \"service-ca-operator-777779d784-t9rgf\" (UID: \"2f9658c9-2fb3-44c1-bb34-62a2e16360bb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9rgf" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.021605 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ed3f40b1-ee95-427b-a6ea-0b412dba932c-socket-dir\") pod \"csi-hostpathplugin-wc8cd\" (UID: \"ed3f40b1-ee95-427b-a6ea-0b412dba932c\") " pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.021622 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ed3f40b1-ee95-427b-a6ea-0b412dba932c-csi-data-dir\") pod \"csi-hostpathplugin-wc8cd\" (UID: \"ed3f40b1-ee95-427b-a6ea-0b412dba932c\") " pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.021694 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf9z4\" (UniqueName: \"kubernetes.io/projected/ed3f40b1-ee95-427b-a6ea-0b412dba932c-kube-api-access-gf9z4\") pod \"csi-hostpathplugin-wc8cd\" (UID: \"ed3f40b1-ee95-427b-a6ea-0b412dba932c\") " pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.021729 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:43 crc kubenswrapper[4578]: E1003 12:53:43.022003 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:43.521987944 +0000 UTC m=+159.320460128 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.024841 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ac7d947f-4847-4809-811d-5fc94e141dd4-webhook-cert\") pod \"packageserver-d55dfcdfc-g5m4p\" (UID: \"ac7d947f-4847-4809-811d-5fc94e141dd4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.025525 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ed3f40b1-ee95-427b-a6ea-0b412dba932c-mountpoint-dir\") pod \"csi-hostpathplugin-wc8cd\" (UID: \"ed3f40b1-ee95-427b-a6ea-0b412dba932c\") " pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.025563 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1c26320a-db2d-4b15-84ff-4b70590a946a-cert\") pod \"ingress-canary-fgghg\" (UID: \"1c26320a-db2d-4b15-84ff-4b70590a946a\") " pod="openshift-ingress-canary/ingress-canary-fgghg" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.025772 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.025823 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02-profile-collector-cert\") pod \"olm-operator-6b444d44fb-dd9jp\" (UID: \"b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.025849 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86765ff0-85f5-462b-9141-0622f547152f-config-volume\") pod \"dns-default-z7c9x\" (UID: \"86765ff0-85f5-462b-9141-0622f547152f\") " pod="openshift-dns/dns-default-z7c9x" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.025902 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46gvn\" (UniqueName: \"kubernetes.io/projected/86765ff0-85f5-462b-9141-0622f547152f-kube-api-access-46gvn\") pod \"dns-default-z7c9x\" (UID: \"86765ff0-85f5-462b-9141-0622f547152f\") " pod="openshift-dns/dns-default-z7c9x" Oct 03 12:53:43 crc kubenswrapper[4578]: E1003 12:53:43.031412 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:43.531398878 +0000 UTC m=+159.329871062 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.035468 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.043909 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.043989 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8h85\" (UniqueName: \"kubernetes.io/projected/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-kube-api-access-f8h85\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.044003 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.044540 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.045311 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bhqv\" (UniqueName: \"kubernetes.io/projected/1c26320a-db2d-4b15-84ff-4b70590a946a-kube-api-access-4bhqv\") pod \"ingress-canary-fgghg\" (UID: \"1c26320a-db2d-4b15-84ff-4b70590a946a\") " pod="openshift-ingress-canary/ingress-canary-fgghg" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.045421 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-registry-certificates\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.047465 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fdfd4313-d05e-4cbd-8520-3f929675c662-secret-volume\") pod \"collect-profiles-29324925-64fg4\" (UID: \"fdfd4313-d05e-4cbd-8520-3f929675c662\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.057493 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-trusted-ca\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.057862 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfvw5\" (UniqueName: \"kubernetes.io/projected/b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02-kube-api-access-gfvw5\") pod \"olm-operator-6b444d44fb-dd9jp\" (UID: \"b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.059156 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02-srv-cert\") pod \"olm-operator-6b444d44fb-dd9jp\" (UID: \"b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.059208 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/86765ff0-85f5-462b-9141-0622f547152f-metrics-tls\") pod \"dns-default-z7c9x\" (UID: \"86765ff0-85f5-462b-9141-0622f547152f\") " pod="openshift-dns/dns-default-z7c9x" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.059318 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trtwx\" (UniqueName: \"kubernetes.io/projected/4ffccf50-3c12-46e8-9445-a5327d57ce2d-kube-api-access-trtwx\") pod \"package-server-manager-789f6589d5-rb78z\" (UID: \"4ffccf50-3c12-46e8-9445-a5327d57ce2d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rb78z" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.060950 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-registry-certificates\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.062342 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4ffccf50-3c12-46e8-9445-a5327d57ce2d-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-rb78z\" (UID: \"4ffccf50-3c12-46e8-9445-a5327d57ce2d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rb78z" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.062376 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fdfd4313-d05e-4cbd-8520-3f929675c662-config-volume\") pod \"collect-profiles-29324925-64fg4\" (UID: \"fdfd4313-d05e-4cbd-8520-3f929675c662\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.062498 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ed3f40b1-ee95-427b-a6ea-0b412dba932c-plugins-dir\") pod \"csi-hostpathplugin-wc8cd\" (UID: \"ed3f40b1-ee95-427b-a6ea-0b412dba932c\") " pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.063486 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ac7d947f-4847-4809-811d-5fc94e141dd4-apiservice-cert\") pod \"packageserver-d55dfcdfc-g5m4p\" (UID: \"ac7d947f-4847-4809-811d-5fc94e141dd4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.063692 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f9658c9-2fb3-44c1-bb34-62a2e16360bb-serving-cert\") pod \"service-ca-operator-777779d784-t9rgf\" (UID: \"2f9658c9-2fb3-44c1-bb34-62a2e16360bb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9rgf" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.066522 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-registry-tls\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.067962 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6528b192-d6c1-46e6-8048-58a76459651a-node-bootstrap-token\") pod \"machine-config-server-rtlhn\" (UID: \"6528b192-d6c1-46e6-8048-58a76459651a\") " pod="openshift-machine-config-operator/machine-config-server-rtlhn" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.075431 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-trusted-ca\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.097567 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-77kj8" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.105586 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-5bv45"] Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.105653 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cl28q"] Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.115401 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-bxmjt"] Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.115436 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b6rr5"] Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.115815 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-wmkdd"] Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.122833 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-registry-tls\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.128427 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8h85\" (UniqueName: \"kubernetes.io/projected/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-kube-api-access-f8h85\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.132355 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5svf7"] Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.141290 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-bound-sa-token\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.181488 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:43 crc kubenswrapper[4578]: E1003 12:53:43.182035 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:43.682012071 +0000 UTC m=+159.480484255 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.183283 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ac7d947f-4847-4809-811d-5fc94e141dd4-webhook-cert\") pod \"packageserver-d55dfcdfc-g5m4p\" (UID: \"ac7d947f-4847-4809-811d-5fc94e141dd4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.183332 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1c26320a-db2d-4b15-84ff-4b70590a946a-cert\") pod \"ingress-canary-fgghg\" (UID: \"1c26320a-db2d-4b15-84ff-4b70590a946a\") " pod="openshift-ingress-canary/ingress-canary-fgghg" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.183900 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ed3f40b1-ee95-427b-a6ea-0b412dba932c-mountpoint-dir\") pod \"csi-hostpathplugin-wc8cd\" (UID: \"ed3f40b1-ee95-427b-a6ea-0b412dba932c\") " pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.183945 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.183964 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02-profile-collector-cert\") pod \"olm-operator-6b444d44fb-dd9jp\" (UID: \"b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.183980 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86765ff0-85f5-462b-9141-0622f547152f-config-volume\") pod \"dns-default-z7c9x\" (UID: \"86765ff0-85f5-462b-9141-0622f547152f\") " pod="openshift-dns/dns-default-z7c9x" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.183998 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46gvn\" (UniqueName: \"kubernetes.io/projected/86765ff0-85f5-462b-9141-0622f547152f-kube-api-access-46gvn\") pod \"dns-default-z7c9x\" (UID: \"86765ff0-85f5-462b-9141-0622f547152f\") " pod="openshift-dns/dns-default-z7c9x" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184031 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bhqv\" (UniqueName: \"kubernetes.io/projected/1c26320a-db2d-4b15-84ff-4b70590a946a-kube-api-access-4bhqv\") pod \"ingress-canary-fgghg\" (UID: \"1c26320a-db2d-4b15-84ff-4b70590a946a\") " pod="openshift-ingress-canary/ingress-canary-fgghg" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184052 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fdfd4313-d05e-4cbd-8520-3f929675c662-secret-volume\") pod \"collect-profiles-29324925-64fg4\" (UID: \"fdfd4313-d05e-4cbd-8520-3f929675c662\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184072 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfvw5\" (UniqueName: \"kubernetes.io/projected/b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02-kube-api-access-gfvw5\") pod \"olm-operator-6b444d44fb-dd9jp\" (UID: \"b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184089 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02-srv-cert\") pod \"olm-operator-6b444d44fb-dd9jp\" (UID: \"b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184105 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/86765ff0-85f5-462b-9141-0622f547152f-metrics-tls\") pod \"dns-default-z7c9x\" (UID: \"86765ff0-85f5-462b-9141-0622f547152f\") " pod="openshift-dns/dns-default-z7c9x" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184125 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trtwx\" (UniqueName: \"kubernetes.io/projected/4ffccf50-3c12-46e8-9445-a5327d57ce2d-kube-api-access-trtwx\") pod \"package-server-manager-789f6589d5-rb78z\" (UID: \"4ffccf50-3c12-46e8-9445-a5327d57ce2d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rb78z" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184151 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4ffccf50-3c12-46e8-9445-a5327d57ce2d-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-rb78z\" (UID: \"4ffccf50-3c12-46e8-9445-a5327d57ce2d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rb78z" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184165 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fdfd4313-d05e-4cbd-8520-3f929675c662-config-volume\") pod \"collect-profiles-29324925-64fg4\" (UID: \"fdfd4313-d05e-4cbd-8520-3f929675c662\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184181 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ed3f40b1-ee95-427b-a6ea-0b412dba932c-plugins-dir\") pod \"csi-hostpathplugin-wc8cd\" (UID: \"ed3f40b1-ee95-427b-a6ea-0b412dba932c\") " pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184201 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ac7d947f-4847-4809-811d-5fc94e141dd4-apiservice-cert\") pod \"packageserver-d55dfcdfc-g5m4p\" (UID: \"ac7d947f-4847-4809-811d-5fc94e141dd4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184216 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f9658c9-2fb3-44c1-bb34-62a2e16360bb-serving-cert\") pod \"service-ca-operator-777779d784-t9rgf\" (UID: \"2f9658c9-2fb3-44c1-bb34-62a2e16360bb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9rgf" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184242 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6528b192-d6c1-46e6-8048-58a76459651a-node-bootstrap-token\") pod \"machine-config-server-rtlhn\" (UID: \"6528b192-d6c1-46e6-8048-58a76459651a\") " pod="openshift-machine-config-operator/machine-config-server-rtlhn" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184261 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cm8v\" (UniqueName: \"kubernetes.io/projected/ac7d947f-4847-4809-811d-5fc94e141dd4-kube-api-access-5cm8v\") pod \"packageserver-d55dfcdfc-g5m4p\" (UID: \"ac7d947f-4847-4809-811d-5fc94e141dd4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184281 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4cdt\" (UniqueName: \"kubernetes.io/projected/fdfd4313-d05e-4cbd-8520-3f929675c662-kube-api-access-x4cdt\") pod \"collect-profiles-29324925-64fg4\" (UID: \"fdfd4313-d05e-4cbd-8520-3f929675c662\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184303 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj2bp\" (UniqueName: \"kubernetes.io/projected/6528b192-d6c1-46e6-8048-58a76459651a-kube-api-access-gj2bp\") pod \"machine-config-server-rtlhn\" (UID: \"6528b192-d6c1-46e6-8048-58a76459651a\") " pod="openshift-machine-config-operator/machine-config-server-rtlhn" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184320 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ed3f40b1-ee95-427b-a6ea-0b412dba932c-registration-dir\") pod \"csi-hostpathplugin-wc8cd\" (UID: \"ed3f40b1-ee95-427b-a6ea-0b412dba932c\") " pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184334 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f9658c9-2fb3-44c1-bb34-62a2e16360bb-config\") pod \"service-ca-operator-777779d784-t9rgf\" (UID: \"2f9658c9-2fb3-44c1-bb34-62a2e16360bb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9rgf" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184352 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6528b192-d6c1-46e6-8048-58a76459651a-certs\") pod \"machine-config-server-rtlhn\" (UID: \"6528b192-d6c1-46e6-8048-58a76459651a\") " pod="openshift-machine-config-operator/machine-config-server-rtlhn" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184369 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ac7d947f-4847-4809-811d-5fc94e141dd4-tmpfs\") pod \"packageserver-d55dfcdfc-g5m4p\" (UID: \"ac7d947f-4847-4809-811d-5fc94e141dd4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184393 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbzl5\" (UniqueName: \"kubernetes.io/projected/2f9658c9-2fb3-44c1-bb34-62a2e16360bb-kube-api-access-pbzl5\") pod \"service-ca-operator-777779d784-t9rgf\" (UID: \"2f9658c9-2fb3-44c1-bb34-62a2e16360bb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9rgf" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184500 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ed3f40b1-ee95-427b-a6ea-0b412dba932c-socket-dir\") pod \"csi-hostpathplugin-wc8cd\" (UID: \"ed3f40b1-ee95-427b-a6ea-0b412dba932c\") " pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184515 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ed3f40b1-ee95-427b-a6ea-0b412dba932c-csi-data-dir\") pod \"csi-hostpathplugin-wc8cd\" (UID: \"ed3f40b1-ee95-427b-a6ea-0b412dba932c\") " pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.184530 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf9z4\" (UniqueName: \"kubernetes.io/projected/ed3f40b1-ee95-427b-a6ea-0b412dba932c-kube-api-access-gf9z4\") pod \"csi-hostpathplugin-wc8cd\" (UID: \"ed3f40b1-ee95-427b-a6ea-0b412dba932c\") " pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.188738 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ed3f40b1-ee95-427b-a6ea-0b412dba932c-mountpoint-dir\") pod \"csi-hostpathplugin-wc8cd\" (UID: \"ed3f40b1-ee95-427b-a6ea-0b412dba932c\") " pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" Oct 03 12:53:43 crc kubenswrapper[4578]: E1003 12:53:43.189043 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:43.689030937 +0000 UTC m=+159.487503121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.199117 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ed3f40b1-ee95-427b-a6ea-0b412dba932c-socket-dir\") pod \"csi-hostpathplugin-wc8cd\" (UID: \"ed3f40b1-ee95-427b-a6ea-0b412dba932c\") " pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.200389 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ac7d947f-4847-4809-811d-5fc94e141dd4-tmpfs\") pod \"packageserver-d55dfcdfc-g5m4p\" (UID: \"ac7d947f-4847-4809-811d-5fc94e141dd4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.204487 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/6528b192-d6c1-46e6-8048-58a76459651a-certs\") pod \"machine-config-server-rtlhn\" (UID: \"6528b192-d6c1-46e6-8048-58a76459651a\") " pod="openshift-machine-config-operator/machine-config-server-rtlhn" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.204638 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ed3f40b1-ee95-427b-a6ea-0b412dba932c-registration-dir\") pod \"csi-hostpathplugin-wc8cd\" (UID: \"ed3f40b1-ee95-427b-a6ea-0b412dba932c\") " pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.206421 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86765ff0-85f5-462b-9141-0622f547152f-config-volume\") pod \"dns-default-z7c9x\" (UID: \"86765ff0-85f5-462b-9141-0622f547152f\") " pod="openshift-dns/dns-default-z7c9x" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.209820 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ed3f40b1-ee95-427b-a6ea-0b412dba932c-csi-data-dir\") pod \"csi-hostpathplugin-wc8cd\" (UID: \"ed3f40b1-ee95-427b-a6ea-0b412dba932c\") " pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.210538 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ac7d947f-4847-4809-811d-5fc94e141dd4-webhook-cert\") pod \"packageserver-d55dfcdfc-g5m4p\" (UID: \"ac7d947f-4847-4809-811d-5fc94e141dd4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.215208 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ed3f40b1-ee95-427b-a6ea-0b412dba932c-plugins-dir\") pod \"csi-hostpathplugin-wc8cd\" (UID: \"ed3f40b1-ee95-427b-a6ea-0b412dba932c\") " pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.213211 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fdfd4313-d05e-4cbd-8520-3f929675c662-config-volume\") pod \"collect-profiles-29324925-64fg4\" (UID: \"fdfd4313-d05e-4cbd-8520-3f929675c662\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.228211 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/6528b192-d6c1-46e6-8048-58a76459651a-node-bootstrap-token\") pod \"machine-config-server-rtlhn\" (UID: \"6528b192-d6c1-46e6-8048-58a76459651a\") " pod="openshift-machine-config-operator/machine-config-server-rtlhn" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.242138 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1c26320a-db2d-4b15-84ff-4b70590a946a-cert\") pod \"ingress-canary-fgghg\" (UID: \"1c26320a-db2d-4b15-84ff-4b70590a946a\") " pod="openshift-ingress-canary/ingress-canary-fgghg" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.243662 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f9658c9-2fb3-44c1-bb34-62a2e16360bb-config\") pod \"service-ca-operator-777779d784-t9rgf\" (UID: \"2f9658c9-2fb3-44c1-bb34-62a2e16360bb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9rgf" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.253938 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ac7d947f-4847-4809-811d-5fc94e141dd4-apiservice-cert\") pod \"packageserver-d55dfcdfc-g5m4p\" (UID: \"ac7d947f-4847-4809-811d-5fc94e141dd4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.256348 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf9z4\" (UniqueName: \"kubernetes.io/projected/ed3f40b1-ee95-427b-a6ea-0b412dba932c-kube-api-access-gf9z4\") pod \"csi-hostpathplugin-wc8cd\" (UID: \"ed3f40b1-ee95-427b-a6ea-0b412dba932c\") " pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.256878 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fdfd4313-d05e-4cbd-8520-3f929675c662-secret-volume\") pod \"collect-profiles-29324925-64fg4\" (UID: \"fdfd4313-d05e-4cbd-8520-3f929675c662\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.257432 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02-profile-collector-cert\") pod \"olm-operator-6b444d44fb-dd9jp\" (UID: \"b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.257834 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbzl5\" (UniqueName: \"kubernetes.io/projected/2f9658c9-2fb3-44c1-bb34-62a2e16360bb-kube-api-access-pbzl5\") pod \"service-ca-operator-777779d784-t9rgf\" (UID: \"2f9658c9-2fb3-44c1-bb34-62a2e16360bb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9rgf" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.258045 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02-srv-cert\") pod \"olm-operator-6b444d44fb-dd9jp\" (UID: \"b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.259802 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/86765ff0-85f5-462b-9141-0622f547152f-metrics-tls\") pod \"dns-default-z7c9x\" (UID: \"86765ff0-85f5-462b-9141-0622f547152f\") " pod="openshift-dns/dns-default-z7c9x" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.261789 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-d2rzv"] Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.262528 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4ffccf50-3c12-46e8-9445-a5327d57ce2d-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-rb78z\" (UID: \"4ffccf50-3c12-46e8-9445-a5327d57ce2d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rb78z" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.263192 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f9658c9-2fb3-44c1-bb34-62a2e16360bb-serving-cert\") pod \"service-ca-operator-777779d784-t9rgf\" (UID: \"2f9658c9-2fb3-44c1-bb34-62a2e16360bb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9rgf" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.264394 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj2bp\" (UniqueName: \"kubernetes.io/projected/6528b192-d6c1-46e6-8048-58a76459651a-kube-api-access-gj2bp\") pod \"machine-config-server-rtlhn\" (UID: \"6528b192-d6c1-46e6-8048-58a76459651a\") " pod="openshift-machine-config-operator/machine-config-server-rtlhn" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.273332 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4cdt\" (UniqueName: \"kubernetes.io/projected/fdfd4313-d05e-4cbd-8520-3f929675c662-kube-api-access-x4cdt\") pod \"collect-profiles-29324925-64fg4\" (UID: \"fdfd4313-d05e-4cbd-8520-3f929675c662\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.285082 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gmbjm"] Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.285451 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:43 crc kubenswrapper[4578]: E1003 12:53:43.285873 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:43.785856178 +0000 UTC m=+159.584328362 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.305590 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cm8v\" (UniqueName: \"kubernetes.io/projected/ac7d947f-4847-4809-811d-5fc94e141dd4-kube-api-access-5cm8v\") pod \"packageserver-d55dfcdfc-g5m4p\" (UID: \"ac7d947f-4847-4809-811d-5fc94e141dd4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.307379 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2cj4m"] Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.316581 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7br6x"] Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.319885 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46gvn\" (UniqueName: \"kubernetes.io/projected/86765ff0-85f5-462b-9141-0622f547152f-kube-api-access-46gvn\") pod \"dns-default-z7c9x\" (UID: \"86765ff0-85f5-462b-9141-0622f547152f\") " pod="openshift-dns/dns-default-z7c9x" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.336252 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bhqv\" (UniqueName: \"kubernetes.io/projected/1c26320a-db2d-4b15-84ff-4b70590a946a-kube-api-access-4bhqv\") pod \"ingress-canary-fgghg\" (UID: \"1c26320a-db2d-4b15-84ff-4b70590a946a\") " pod="openshift-ingress-canary/ingress-canary-fgghg" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.345155 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trtwx\" (UniqueName: \"kubernetes.io/projected/4ffccf50-3c12-46e8-9445-a5327d57ce2d-kube-api-access-trtwx\") pod \"package-server-manager-789f6589d5-rb78z\" (UID: \"4ffccf50-3c12-46e8-9445-a5327d57ce2d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rb78z" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.371374 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfvw5\" (UniqueName: \"kubernetes.io/projected/b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02-kube-api-access-gfvw5\") pod \"olm-operator-6b444d44fb-dd9jp\" (UID: \"b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp" Oct 03 12:53:43 crc kubenswrapper[4578]: W1003 12:53:43.383656 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod185f4f54_9f50_4017_a7fc_7c7821338442.slice/crio-32f10046bb13b23b0124992fd4dce3b3197ff33fd27ebbbfa141a973b47e41a2 WatchSource:0}: Error finding container 32f10046bb13b23b0124992fd4dce3b3197ff33fd27ebbbfa141a973b47e41a2: Status 404 returned error can't find the container with id 32f10046bb13b23b0124992fd4dce3b3197ff33fd27ebbbfa141a973b47e41a2 Oct 03 12:53:43 crc kubenswrapper[4578]: W1003 12:53:43.386822 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89b4a9e4_3b97_4ebc_850e_8995efbbaf7e.slice/crio-6d3abbb8b3f38e30706c3b831a4b9960371b460deb29c40ac145ecc356e21e58 WatchSource:0}: Error finding container 6d3abbb8b3f38e30706c3b831a4b9960371b460deb29c40ac145ecc356e21e58: Status 404 returned error can't find the container with id 6d3abbb8b3f38e30706c3b831a4b9960371b460deb29c40ac145ecc356e21e58 Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.387528 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:43 crc kubenswrapper[4578]: E1003 12:53:43.387953 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:43.887900557 +0000 UTC m=+159.686372741 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.422739 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.438478 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8flns"] Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.450211 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.462030 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.472108 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rb78z" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.480082 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9rgf" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.488323 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:43 crc kubenswrapper[4578]: E1003 12:53:43.488542 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:43.98852753 +0000 UTC m=+159.786999714 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.488623 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-z7c9x" Oct 03 12:53:43 crc kubenswrapper[4578]: W1003 12:53:43.492119 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09d82c69_039a_4e2a_bcd7_dad26f5c8486.slice/crio-328543c0eef4437eda8fbc13756dfd618dadac7f05eac12ba1c105b3b10e1654 WatchSource:0}: Error finding container 328543c0eef4437eda8fbc13756dfd618dadac7f05eac12ba1c105b3b10e1654: Status 404 returned error can't find the container with id 328543c0eef4437eda8fbc13756dfd618dadac7f05eac12ba1c105b3b10e1654 Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.497863 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-fgghg" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.506881 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-rtlhn" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.526034 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-8htrd"] Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.526202 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.549942 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-l2z5f"] Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.590490 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:43 crc kubenswrapper[4578]: E1003 12:53:43.590891 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:44.090878359 +0000 UTC m=+159.889350543 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:43 crc kubenswrapper[4578]: W1003 12:53:43.619362 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc48e4a1e_a25a_49d8_aa80_bd1e7425ae33.slice/crio-f22580ca60e3ac786759960c2d8b1619219cb898fcffba25774f991fe710fc51 WatchSource:0}: Error finding container f22580ca60e3ac786759960c2d8b1619219cb898fcffba25774f991fe710fc51: Status 404 returned error can't find the container with id f22580ca60e3ac786759960c2d8b1619219cb898fcffba25774f991fe710fc51 Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.626279 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7kdcr"] Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.635592 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-r9hlj"] Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.650337 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sqbkd"] Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.666134 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hpc5d"] Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.671529 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz"] Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.692994 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:43 crc kubenswrapper[4578]: E1003 12:53:43.693406 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:44.193386493 +0000 UTC m=+159.991858677 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.709315 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4qsl8"] Oct 03 12:53:43 crc kubenswrapper[4578]: W1003 12:53:43.725015 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92275d4e_f157_49a7_b292_2cd46bc23c11.slice/crio-058f5a4ff23b92ea5395cf7dd306672d26ec043ac0bbec57bc46ccd0ff38c594 WatchSource:0}: Error finding container 058f5a4ff23b92ea5395cf7dd306672d26ec043ac0bbec57bc46ccd0ff38c594: Status 404 returned error can't find the container with id 058f5a4ff23b92ea5395cf7dd306672d26ec043ac0bbec57bc46ccd0ff38c594 Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.775669 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8flns" event={"ID":"09d82c69-039a-4e2a-bcd7-dad26f5c8486","Type":"ContainerStarted","Data":"328543c0eef4437eda8fbc13756dfd618dadac7f05eac12ba1c105b3b10e1654"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.787402 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b"] Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.788564 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zsb7g" event={"ID":"2f691127-6990-46bd-afa7-d6f3f7eaa329","Type":"ContainerStarted","Data":"cd327fc6d984b60c85562fac9b5e254029a86325c99cd5b756547af472bf4337"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.788602 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-zsb7g" event={"ID":"2f691127-6990-46bd-afa7-d6f3f7eaa329","Type":"ContainerStarted","Data":"f74c28d05998f10db66bc0c5f37fd7a4ff8b97e40dde9de77129c53e6de48145"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.796719 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:43 crc kubenswrapper[4578]: E1003 12:53:43.797476 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:44.297462867 +0000 UTC m=+160.095935051 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.802943 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" event={"ID":"5a1d0b1c-9bae-483f-886d-bf4acffee7d9","Type":"ContainerStarted","Data":"b1dda7af6c7a57e9ca017010f0339fcf1e4b44c91b868d9c9f7093414c9538b5"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.803982 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.809981 4578 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-t9q7n container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.810018 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" podUID="5a1d0b1c-9bae-483f-886d-bf4acffee7d9" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.812732 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-bxmjt" event={"ID":"9a11b49f-69d2-4d9d-a451-7b69b797da20","Type":"ContainerStarted","Data":"6d8424d54446a4f665bf7b8ae77e48a7040e61f1f2a3d0734af7687ef8f39334"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.818310 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-4s8fg" event={"ID":"dea78529-724d-461e-a64c-ed4858ff983c","Type":"ContainerStarted","Data":"e0450fd99b27ff12f025ffc0328c508528cb292996cf916f2262d875b8c0a2c8"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.830371 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8htrd" event={"ID":"c48e4a1e-a25a-49d8-aa80-bd1e7425ae33","Type":"ContainerStarted","Data":"f22580ca60e3ac786759960c2d8b1619219cb898fcffba25774f991fe710fc51"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.833812 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mt2gv" event={"ID":"1a194512-24b8-43b7-89f9-db084d65ee0a","Type":"ContainerStarted","Data":"18e05669918bb17d392dfc0bc973ee99bef6aa437ae0ddfa591edc0449bc9119"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.836768 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2cj4m" event={"ID":"185f4f54-9f50-4017-a7fc-7c7821338442","Type":"ContainerStarted","Data":"32f10046bb13b23b0124992fd4dce3b3197ff33fd27ebbbfa141a973b47e41a2"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.841741 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" event={"ID":"2d27f0e8-6b63-4128-b886-7f0018a6a18b","Type":"ContainerStarted","Data":"fdfc4370051fd90b202ffd6f7957abf574b0b2f4fa0ef047271769ee59f7ebd9"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.843886 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-d2rzv" event={"ID":"de78c746-e145-48ba-8601-a3ff2c6b2b62","Type":"ContainerStarted","Data":"9976598c3b87638be7f4b8763cf329f69d41e020d9e7c125b84932fd687b35d8"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.845102 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5svf7" event={"ID":"e9ffa547-6ab2-4884-8433-d80fff56b441","Type":"ContainerStarted","Data":"140a47461397eb6e368f5e13bbf1ee46f8d3c7b123ca60ce24cef77fb2f46db1"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.848218 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" event={"ID":"ad129323-bfad-47a1-84a5-7b51025f5d0d","Type":"ContainerStarted","Data":"fa8d1b7e9f65e4a12809db42e407577c589b386e5052ad20d2e2f975d6681e05"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.853068 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-5bv45" event={"ID":"6c98467a-2f3b-4329-8f21-9492b087b5d8","Type":"ContainerStarted","Data":"fc94958ea7b47e07b36dcde808dd632582f245e2dc4c019b9a8b74fbc5198116"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.853998 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-l2z5f" event={"ID":"92275d4e-f157-49a7-b292-2cd46bc23c11","Type":"ContainerStarted","Data":"058f5a4ff23b92ea5395cf7dd306672d26ec043ac0bbec57bc46ccd0ff38c594"} Oct 03 12:53:43 crc kubenswrapper[4578]: W1003 12:53:43.857387 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod959dff27_2ebe_417a_b5be_3e814dff3797.slice/crio-ba9cfbe0c98f0886c92b0e92d7031336186381e58fe151f641af430e44416666 WatchSource:0}: Error finding container ba9cfbe0c98f0886c92b0e92d7031336186381e58fe151f641af430e44416666: Status 404 returned error can't find the container with id ba9cfbe0c98f0886c92b0e92d7031336186381e58fe151f641af430e44416666 Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.862420 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cxghf" event={"ID":"bc81f2fa-5ee7-4034-81d3-a11639762ac9","Type":"ContainerStarted","Data":"321d6fc2c75b4403014a4bef494d1a278b32e4d31b8d8cca1a744afed16eb557"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.863617 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" event={"ID":"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e","Type":"ContainerStarted","Data":"6d3abbb8b3f38e30706c3b831a4b9960371b460deb29c40ac145ecc356e21e58"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.865014 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b6rr5" event={"ID":"6b207ee8-d007-408b-8519-b6fa02401fd0","Type":"ContainerStarted","Data":"feb069dcc93f8893540e491d045b4b85f2a2859b54e46d90da6a386272a859ca"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.865045 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b6rr5" event={"ID":"6b207ee8-d007-408b-8519-b6fa02401fd0","Type":"ContainerStarted","Data":"a6061523d4255e89489ac9986280930ab32694d43f074909788f74ee1ea22534"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.867832 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-q8kdg" event={"ID":"59bf594f-62b6-4a2f-a50d-30b5d8357133","Type":"ContainerStarted","Data":"2eb7bc1615e43ea4988b98f37ea586d165987ed2ced6cdd0c7c84ba9829d9ea9"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.868477 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" event={"ID":"b2ee8da6-43e0-496d-aea1-2c7849305419","Type":"ContainerStarted","Data":"a0e3082a18351f620c4121d58b0d8d6d395a7706e4c6e90ebf3d8ced17b5fcb6"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.869424 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gmbjm" event={"ID":"1a3d0383-95ba-457c-af8c-0eebe774e2ea","Type":"ContainerStarted","Data":"35e25f037bd7f016bc658315ef32d8f2375b96e6fdcdf2ff00e9be31f5321e45"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.871514 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" event={"ID":"bec2b13f-9bf4-4378-8a9c-56e61397071a","Type":"ContainerStarted","Data":"d3c5bfafc3ae90bcc6c82cd7d68a635558fbe0f43cb37e714567ec630bcbee30"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.873192 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ns26l" event={"ID":"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6","Type":"ContainerStarted","Data":"abb6d94d578781509f9906a23838695125978d11ac4cb25e065f03fe458ae071"} Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.899956 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:43 crc kubenswrapper[4578]: E1003 12:53:43.900986 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:44.400966942 +0000 UTC m=+160.199439146 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.911024 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:43 crc kubenswrapper[4578]: E1003 12:53:43.913100 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:44.413083563 +0000 UTC m=+160.211555737 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:43 crc kubenswrapper[4578]: I1003 12:53:43.917874 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-m68m5"] Oct 03 12:53:43 crc kubenswrapper[4578]: W1003 12:53:43.969883 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5581f00e_5f39_454d_843d_9dbf6d2cbb0b.slice/crio-d55c8562f345747a21f8d201d8ae862f701b3dd9342ba9e2de0f7e18cdcbe6bb WatchSource:0}: Error finding container d55c8562f345747a21f8d201d8ae862f701b3dd9342ba9e2de0f7e18cdcbe6bb: Status 404 returned error can't find the container with id d55c8562f345747a21f8d201d8ae862f701b3dd9342ba9e2de0f7e18cdcbe6bb Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.015759 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:44 crc kubenswrapper[4578]: E1003 12:53:44.016596 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:44.516571378 +0000 UTC m=+160.315043562 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.016911 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:44 crc kubenswrapper[4578]: E1003 12:53:44.019051 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:44.519027027 +0000 UTC m=+160.317499211 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.073903 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rb78z"] Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.081086 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xctz5"] Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.106056 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-77kj8"] Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.123733 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:44 crc kubenswrapper[4578]: E1003 12:53:44.129310 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:44.62926527 +0000 UTC m=+160.427737454 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.139890 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-b6rr5" podStartSLOduration=133.139868491 podStartE2EDuration="2m13.139868491s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:44.129843899 +0000 UTC m=+159.928316083" watchObservedRunningTime="2025-10-03 12:53:44.139868491 +0000 UTC m=+159.938340675" Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.209601 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-4s8fg" podStartSLOduration=134.209467435 podStartE2EDuration="2m14.209467435s" podCreationTimestamp="2025-10-03 12:51:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:44.209283089 +0000 UTC m=+160.007755273" watchObservedRunningTime="2025-10-03 12:53:44.209467435 +0000 UTC m=+160.007939619" Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.225065 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:44 crc kubenswrapper[4578]: E1003 12:53:44.225707 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:44.725695868 +0000 UTC m=+160.524168052 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.362227 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:44 crc kubenswrapper[4578]: E1003 12:53:44.363781 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:44.863752918 +0000 UTC m=+160.662225122 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.421901 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p"] Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.466706 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:44 crc kubenswrapper[4578]: E1003 12:53:44.474058 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:44.974042582 +0000 UTC m=+160.772514766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.498692 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-wc8cd"] Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.498747 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-zsb7g" podStartSLOduration=133.498729428 podStartE2EDuration="2m13.498729428s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:44.49013654 +0000 UTC m=+160.288608724" watchObservedRunningTime="2025-10-03 12:53:44.498729428 +0000 UTC m=+160.297201612" Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.568370 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:44 crc kubenswrapper[4578]: E1003 12:53:44.568760 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:45.068741833 +0000 UTC m=+160.867214017 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.672727 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:44 crc kubenswrapper[4578]: E1003 12:53:44.673082 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:45.173068276 +0000 UTC m=+160.971540460 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.773718 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:44 crc kubenswrapper[4578]: E1003 12:53:44.774414 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:45.274386292 +0000 UTC m=+161.072858476 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.780524 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" podStartSLOduration=133.780502378 podStartE2EDuration="2m13.780502378s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:44.729214226 +0000 UTC m=+160.527686410" watchObservedRunningTime="2025-10-03 12:53:44.780502378 +0000 UTC m=+160.578974562" Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.810172 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mt2gv" podStartSLOduration=134.810151184 podStartE2EDuration="2m14.810151184s" podCreationTimestamp="2025-10-03 12:51:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:44.809300757 +0000 UTC m=+160.607772951" watchObservedRunningTime="2025-10-03 12:53:44.810151184 +0000 UTC m=+160.608623378" Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.857775 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4"] Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.862994 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-z7c9x"] Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.875566 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:44 crc kubenswrapper[4578]: E1003 12:53:44.875978 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:45.375950754 +0000 UTC m=+161.174422938 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.927183 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m68m5" event={"ID":"fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc","Type":"ContainerStarted","Data":"48ac3cd1d54bab4f3f052c10e989ae2b6a31eb3a1fcffdc6702d3c1db3b5307e"} Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.976556 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:44 crc kubenswrapper[4578]: E1003 12:53:44.977317 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:45.477303701 +0000 UTC m=+161.275775885 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:44 crc kubenswrapper[4578]: I1003 12:53:44.990797 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-fgghg"] Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.032902 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp"] Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.057751 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" event={"ID":"b2ee8da6-43e0-496d-aea1-2c7849305419","Type":"ContainerStarted","Data":"543baf20b344ae188ff98b93c23363072c8aafa779ad8e8e546be787bb6c7df1"} Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.065418 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.095106 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:45 crc kubenswrapper[4578]: E1003 12:53:45.095571 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:45.595555852 +0000 UTC m=+161.394028036 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.099133 4578 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-hh7tq container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.099209 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" podUID="b2ee8da6-43e0-496d-aea1-2c7849305419" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.102857 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2cj4m" event={"ID":"185f4f54-9f50-4017-a7fc-7c7821338442","Type":"ContainerStarted","Data":"e49c217829be7c4318874a7da0c4f439c335e89b00d8271848ef95d98fba4174"} Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.146805 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-r9hlj" event={"ID":"f78b5fb4-0a61-4222-928f-2383bac97358","Type":"ContainerStarted","Data":"c6368916a8ce04edb1ec0bfab4581334749b90692f408c08c2af7ef0fa0f52fe"} Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.178280 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ns26l" event={"ID":"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6","Type":"ContainerStarted","Data":"841a71102e1dc33e360fa2d6d0e2bdb8b21bed4852a9dbaa4e6c2519e39b7a0f"} Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.198172 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:45 crc kubenswrapper[4578]: E1003 12:53:45.198640 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:45.698605474 +0000 UTC m=+161.497077668 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.198750 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:45 crc kubenswrapper[4578]: E1003 12:53:45.200049 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:45.7000346 +0000 UTC m=+161.498506784 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.200991 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-t9rgf"] Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.224923 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-d2rzv" event={"ID":"de78c746-e145-48ba-8601-a3ff2c6b2b62","Type":"ContainerStarted","Data":"26167cd8f8d93e7c8d3ecce6a842ff40e7741aa64ba5c845485a43cdd0e77153"} Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.225275 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-d2rzv" Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.226763 4578 patch_prober.go:28] interesting pod/downloads-7954f5f757-d2rzv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.227068 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-d2rzv" podUID="de78c746-e145-48ba-8601-a3ff2c6b2b62" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.230543 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8flns" event={"ID":"09d82c69-039a-4e2a-bcd7-dad26f5c8486","Type":"ContainerStarted","Data":"9a25bbd04c4a86646c05069e5393baae4e11f3c560efba0277eb739c1126a9c9"} Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.245935 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" event={"ID":"ed3f40b1-ee95-427b-a6ea-0b412dba932c","Type":"ContainerStarted","Data":"fa87fa398d4394a557293745ec6bb90307685a581d7c728f9a53b58e380b7616"} Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.267937 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-bxmjt" event={"ID":"9a11b49f-69d2-4d9d-a451-7b69b797da20","Type":"ContainerStarted","Data":"debe9bbddd66ff70afdfb272ba06b0eb710b248d79eba91e40807e4994391dbd"} Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.268842 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-bxmjt" Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.277035 4578 patch_prober.go:28] interesting pod/console-operator-58897d9998-bxmjt container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/readyz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.277080 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-bxmjt" podUID="9a11b49f-69d2-4d9d-a451-7b69b797da20" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.17:8443/readyz\": dial tcp 10.217.0.17:8443: connect: connection refused" Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.295758 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" event={"ID":"4ef05574-4924-4a73-b0ba-0e632c96410a","Type":"ContainerStarted","Data":"4ad23a827a231477fdb77ba6ee3f420fc6df78c85d1f747cf3c87cf7c9d7423e"} Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.301808 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:45 crc kubenswrapper[4578]: E1003 12:53:45.302961 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:45.802936975 +0000 UTC m=+161.601409209 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.353762 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5svf7" event={"ID":"e9ffa547-6ab2-4884-8433-d80fff56b441","Type":"ContainerStarted","Data":"2aa266a9de0c059636ad635dcf5de19875da1bbae254a2abdbb8ba557e3adf30"} Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.360646 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7kdcr" event={"ID":"ec125663-8341-427e-917c-758bbeae4251","Type":"ContainerStarted","Data":"537f223e5ac112fc8608e3f96cb03718068b8ef59e46c4a298dafc7f2c875276"} Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.394872 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-rtlhn" event={"ID":"6528b192-d6c1-46e6-8048-58a76459651a","Type":"ContainerStarted","Data":"9e9280e057a55af27165677a56596046e728f7c1162ef49d5cada40957723cda"} Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.405152 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:45 crc kubenswrapper[4578]: E1003 12:53:45.407190 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:45.907175425 +0000 UTC m=+161.705647609 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.419866 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4qsl8" event={"ID":"959dff27-2ebe-417a-b5be-3e814dff3797","Type":"ContainerStarted","Data":"ba9cfbe0c98f0886c92b0e92d7031336186381e58fe151f641af430e44416666"} Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.440677 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xctz5" event={"ID":"db43e12f-df81-44cf-9220-2f75d5ecc70b","Type":"ContainerStarted","Data":"5f69c131eb9f30d65312c42e4004382600c9fd69e4470858d89ca7e0bddbce28"} Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.443523 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hpc5d" event={"ID":"6f3f593c-f784-4f1c-8fb3-016275c38649","Type":"ContainerStarted","Data":"d7c5539eb7c213d95b6f9579d15f722d53be3ab3c06358acd65093c07d9767aa"} Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.461034 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" event={"ID":"2d27f0e8-6b63-4128-b886-7f0018a6a18b","Type":"ContainerStarted","Data":"a8374726ba6a166e982a6876dc4117435c07af958f667348f1d15324abeccaf0"} Oct 03 12:53:45 crc kubenswrapper[4578]: W1003 12:53:45.476953 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f9658c9_2fb3_44c1_bb34_62a2e16360bb.slice/crio-0854dc9dc852cec7bffa0a79f6331a08faeddf4f61a5f2950908a89cfd526544 WatchSource:0}: Error finding container 0854dc9dc852cec7bffa0a79f6331a08faeddf4f61a5f2950908a89cfd526544: Status 404 returned error can't find the container with id 0854dc9dc852cec7bffa0a79f6331a08faeddf4f61a5f2950908a89cfd526544 Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.479574 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-77kj8" event={"ID":"7d9d9160-dd25-4640-a84b-ee9558be3b88","Type":"ContainerStarted","Data":"4913019ed9debdc141593be04568c77f517dc93f250085e21b765663a1f052f9"} Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.485838 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" event={"ID":"ac7d947f-4847-4809-811d-5fc94e141dd4","Type":"ContainerStarted","Data":"d89c53435e64d11d2c55650dd09dfe1b4c70495f9a89e751feaa85e587c6cc45"} Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.494720 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b" event={"ID":"5581f00e-5f39-454d-843d-9dbf6d2cbb0b","Type":"ContainerStarted","Data":"d55c8562f345747a21f8d201d8ae862f701b3dd9342ba9e2de0f7e18cdcbe6bb"} Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.495871 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz" event={"ID":"ee3b2702-a354-4921-a66d-650bedd0c1b2","Type":"ContainerStarted","Data":"46f18269feaea6341e630b6d82ad1a1636cf981991f794b2fea82eaea0637fca"} Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.497514 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rb78z" event={"ID":"4ffccf50-3c12-46e8-9445-a5327d57ce2d","Type":"ContainerStarted","Data":"e82b164e17d5d1f9040c65247efd2e11c411affcfecc5660ff6edb7f984273c3"} Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.506280 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:45 crc kubenswrapper[4578]: E1003 12:53:45.506560 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:46.006546459 +0000 UTC m=+161.805018643 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.510491 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.524185 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-cl28q" podStartSLOduration=134.524166956 podStartE2EDuration="2m14.524166956s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:45.522871084 +0000 UTC m=+161.321343268" watchObservedRunningTime="2025-10-03 12:53:45.524166956 +0000 UTC m=+161.322639140" Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.618916 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-d2rzv" podStartSLOduration=134.618890048 podStartE2EDuration="2m14.618890048s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:45.612597895 +0000 UTC m=+161.411070079" watchObservedRunningTime="2025-10-03 12:53:45.618890048 +0000 UTC m=+161.417362232" Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.620387 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:45 crc kubenswrapper[4578]: E1003 12:53:45.620662 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:46.120621194 +0000 UTC m=+161.919093388 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.673027 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-q8kdg" podStartSLOduration=134.672999683 podStartE2EDuration="2m14.672999683s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:45.654547058 +0000 UTC m=+161.453019252" watchObservedRunningTime="2025-10-03 12:53:45.672999683 +0000 UTC m=+161.471471867" Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.706770 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" podStartSLOduration=134.70674654 podStartE2EDuration="2m14.70674654s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:45.697226164 +0000 UTC m=+161.495698348" watchObservedRunningTime="2025-10-03 12:53:45.70674654 +0000 UTC m=+161.505218714" Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.745674 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:45 crc kubenswrapper[4578]: E1003 12:53:45.746446 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:46.2464319 +0000 UTC m=+162.044904084 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.767826 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2cj4m" podStartSLOduration=135.767808948 podStartE2EDuration="2m15.767808948s" podCreationTimestamp="2025-10-03 12:51:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:45.744965022 +0000 UTC m=+161.543437206" watchObservedRunningTime="2025-10-03 12:53:45.767808948 +0000 UTC m=+161.566281132" Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.769290 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5svf7" podStartSLOduration=134.769283876 podStartE2EDuration="2m14.769283876s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:45.765796793 +0000 UTC m=+161.564268977" watchObservedRunningTime="2025-10-03 12:53:45.769283876 +0000 UTC m=+161.567756060" Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.847603 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:45 crc kubenswrapper[4578]: E1003 12:53:45.847985 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:46.347973762 +0000 UTC m=+162.146445946 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:45 crc kubenswrapper[4578]: I1003 12:53:45.949398 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:45 crc kubenswrapper[4578]: E1003 12:53:45.952731 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:46.452706467 +0000 UTC m=+162.251178651 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.011922 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-ns26l" podStartSLOduration=135.011905785 podStartE2EDuration="2m15.011905785s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:45.948905225 +0000 UTC m=+161.747377409" watchObservedRunningTime="2025-10-03 12:53:46.011905785 +0000 UTC m=+161.810377969" Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.012016 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-bxmjt" podStartSLOduration=135.012013039 podStartE2EDuration="2m15.012013039s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:46.011020117 +0000 UTC m=+161.809492291" watchObservedRunningTime="2025-10-03 12:53:46.012013039 +0000 UTC m=+161.810485213" Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.046492 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.046865 4578 patch_prober.go:28] interesting pod/router-default-5444994796-q8kdg container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.046898 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q8kdg" podUID="59bf594f-62b6-4a2f-a50d-30b5d8357133" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.053388 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:46 crc kubenswrapper[4578]: E1003 12:53:46.053721 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:46.553707552 +0000 UTC m=+162.352179736 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.154143 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:46 crc kubenswrapper[4578]: E1003 12:53:46.154751 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:46.654731268 +0000 UTC m=+162.453203452 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.255481 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:46 crc kubenswrapper[4578]: E1003 12:53:46.255826 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:46.755814646 +0000 UTC m=+162.554286830 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.358359 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:46 crc kubenswrapper[4578]: E1003 12:53:46.358742 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:46.858718813 +0000 UTC m=+162.657191017 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.459543 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:46 crc kubenswrapper[4578]: E1003 12:53:46.459944 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:46.959924364 +0000 UTC m=+162.758396618 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.552805 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-r9hlj" event={"ID":"f78b5fb4-0a61-4222-928f-2383bac97358","Type":"ContainerStarted","Data":"5aaa065afa8af40bcbfab1ac990f9ed2e32573d6e9a4ca25083f76cac702b8e4"} Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.558829 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-z7c9x" event={"ID":"86765ff0-85f5-462b-9141-0622f547152f","Type":"ContainerStarted","Data":"1c7e176f00662772c16850fada46f8d63da54093d6df044e0bda500d548a7d42"} Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.562056 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:46 crc kubenswrapper[4578]: E1003 12:53:46.562567 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:47.062424808 +0000 UTC m=+162.860896992 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.574754 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m68m5" event={"ID":"fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc","Type":"ContainerStarted","Data":"d61f78b7e81bf7441c5c86dae8102b17ddbd252592d16a9cbbd6517df4a39c6c"} Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.630052 4578 generic.go:334] "Generic (PLEG): container finished" podID="c48e4a1e-a25a-49d8-aa80-bd1e7425ae33" containerID="1e706ff4d1673b1fbc715ba54b552a570b9a687f26260b5c20e81db7e6808ab2" exitCode=0 Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.630289 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8htrd" event={"ID":"c48e4a1e-a25a-49d8-aa80-bd1e7425ae33","Type":"ContainerDied","Data":"1e706ff4d1673b1fbc715ba54b552a570b9a687f26260b5c20e81db7e6808ab2"} Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.634993 4578 generic.go:334] "Generic (PLEG): container finished" podID="ad129323-bfad-47a1-84a5-7b51025f5d0d" containerID="12343d79fb8363c55d46fefb2e6529fcaa1211ecb7f269c32d6c851ab65339a7" exitCode=0 Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.635058 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" event={"ID":"ad129323-bfad-47a1-84a5-7b51025f5d0d","Type":"ContainerDied","Data":"12343d79fb8363c55d46fefb2e6529fcaa1211ecb7f269c32d6c851ab65339a7"} Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.636083 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-fgghg" event={"ID":"1c26320a-db2d-4b15-84ff-4b70590a946a","Type":"ContainerStarted","Data":"45043efcca1a0405b2e1eed5167eefb439a68d94e652133cff273cb05b5be95b"} Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.655945 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8flns" event={"ID":"09d82c69-039a-4e2a-bcd7-dad26f5c8486","Type":"ContainerStarted","Data":"1406664c115eba1ef3c3ef97097d9ee7c63398d1a3ba19d414740085e2428c1d"} Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.663973 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:46 crc kubenswrapper[4578]: E1003 12:53:46.666481 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:47.166465722 +0000 UTC m=+162.964937906 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.690899 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-fgghg" podStartSLOduration=6.690878448 podStartE2EDuration="6.690878448s" podCreationTimestamp="2025-10-03 12:53:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:46.689351189 +0000 UTC m=+162.487823363" watchObservedRunningTime="2025-10-03 12:53:46.690878448 +0000 UTC m=+162.489350632" Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.696956 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp" event={"ID":"b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02","Type":"ContainerStarted","Data":"cfde734c526baae8d2f6412cd932ca01104c10f0558b9ccabe544a5700f9f0f6"} Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.698031 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp" Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.714471 4578 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-dd9jp container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:8443/healthz\": dial tcp 10.217.0.42:8443: connect: connection refused" start-of-body= Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.714544 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp" podUID="b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.42:8443/healthz\": dial tcp 10.217.0.42:8443: connect: connection refused" Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.736138 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8flns" podStartSLOduration=135.736118296 podStartE2EDuration="2m15.736118296s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:46.724590974 +0000 UTC m=+162.523063178" watchObservedRunningTime="2025-10-03 12:53:46.736118296 +0000 UTC m=+162.534590480" Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.753818 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gmbjm" event={"ID":"1a3d0383-95ba-457c-af8c-0eebe774e2ea","Type":"ContainerStarted","Data":"b702d0e0d4c2be2775894873f5c4bd36c773f4fff878235ed9459ab2dfeb0d70"} Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.771297 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:46 crc kubenswrapper[4578]: E1003 12:53:46.772684 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:47.272657503 +0000 UTC m=+163.071129687 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.836511 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp" podStartSLOduration=135.836497542 podStartE2EDuration="2m15.836497542s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:46.794931681 +0000 UTC m=+162.593403875" watchObservedRunningTime="2025-10-03 12:53:46.836497542 +0000 UTC m=+162.634969726" Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.853997 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hpc5d" event={"ID":"6f3f593c-f784-4f1c-8fb3-016275c38649","Type":"ContainerStarted","Data":"49759b5ba0354383b0aaff04ec719e02e1f26ddd2fcdd860f295b6a428e5a06e"} Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.874456 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:46 crc kubenswrapper[4578]: E1003 12:53:46.876535 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:47.376522841 +0000 UTC m=+163.174995025 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.926044 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gmbjm" podStartSLOduration=135.926019136 podStartE2EDuration="2m15.926019136s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:46.83955608 +0000 UTC m=+162.638028254" watchObservedRunningTime="2025-10-03 12:53:46.926019136 +0000 UTC m=+162.724491320" Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.926592 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hpc5d" podStartSLOduration=135.926587034 podStartE2EDuration="2m15.926587034s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:46.914057911 +0000 UTC m=+162.712530095" watchObservedRunningTime="2025-10-03 12:53:46.926587034 +0000 UTC m=+162.725059218" Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.955085 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-q8kdg" event={"ID":"59bf594f-62b6-4a2f-a50d-30b5d8357133","Type":"ContainerStarted","Data":"a5aa795b51e39364076fca2a059ea062186c05fef6dd3ea3cf0cd54447c65957"} Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.955397 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-l2z5f" event={"ID":"92275d4e-f157-49a7-b292-2cd46bc23c11","Type":"ContainerStarted","Data":"b73ede4efb4ba3874c66c728ff103530d8902f43d1c5c48344bb7e66e3aeb058"} Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.978934 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:46 crc kubenswrapper[4578]: E1003 12:53:46.980007 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:47.479992726 +0000 UTC m=+163.278464910 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:46 crc kubenswrapper[4578]: I1003 12:53:46.994679 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-77kj8" event={"ID":"7d9d9160-dd25-4640-a84b-ee9558be3b88","Type":"ContainerStarted","Data":"a90ed99cc6eddba5cf158eb6546083717b645be9db21bec54d88437abfc9cb56"} Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.029896 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rb78z" event={"ID":"4ffccf50-3c12-46e8-9445-a5327d57ce2d","Type":"ContainerStarted","Data":"5b175c99b52a1f792a715a0ebd62fe19f59cf5055aa68c83b7d4b99f30244240"} Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.029955 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rb78z" event={"ID":"4ffccf50-3c12-46e8-9445-a5327d57ce2d","Type":"ContainerStarted","Data":"654de47064e1a85fc93964c973a263a7fc6c908bac401155f99e360656b2154d"} Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.030812 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rb78z" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.052712 4578 patch_prober.go:28] interesting pod/router-default-5444994796-q8kdg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 12:53:47 crc kubenswrapper[4578]: [-]has-synced failed: reason withheld Oct 03 12:53:47 crc kubenswrapper[4578]: [+]process-running ok Oct 03 12:53:47 crc kubenswrapper[4578]: healthz check failed Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.052762 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q8kdg" podUID="59bf594f-62b6-4a2f-a50d-30b5d8357133" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.065543 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4" event={"ID":"fdfd4313-d05e-4cbd-8520-3f929675c662","Type":"ContainerStarted","Data":"d48734357a36fe70d9e2e9cd7a2761d3b9aab129eb319e996fb8e83858c1d90d"} Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.071184 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" event={"ID":"4ef05574-4924-4a73-b0ba-0e632c96410a","Type":"ContainerStarted","Data":"439daae67e164ac422ee4754f63691531024d81e693b41b096aecd99ecedc43d"} Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.071815 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.084314 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-l2z5f" podStartSLOduration=136.084282127 podStartE2EDuration="2m16.084282127s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:46.974448047 +0000 UTC m=+162.772920221" watchObservedRunningTime="2025-10-03 12:53:47.084282127 +0000 UTC m=+162.882754311" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.084530 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:47 crc kubenswrapper[4578]: E1003 12:53:47.084881 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:47.584865056 +0000 UTC m=+163.383337240 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.091819 4578 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-sqbkd container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.091890 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" podUID="4ef05574-4924-4a73-b0ba-0e632c96410a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.100480 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7kdcr" event={"ID":"ec125663-8341-427e-917c-758bbeae4251","Type":"ContainerStarted","Data":"a779b147daee8b7960cc0fd7d7ee4885331688c159fe10b4f197dd2fc54a3b5e"} Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.110532 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz" event={"ID":"ee3b2702-a354-4921-a66d-650bedd0c1b2","Type":"ContainerStarted","Data":"37b68e83d0d6392bcd962a87afd29e198dcc2df0ab9862ef2fef61edae11a647"} Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.111312 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.128251 4578 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-l9fbz container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.128299 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz" podUID="ee3b2702-a354-4921-a66d-650bedd0c1b2" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.132272 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" event={"ID":"ac7d947f-4847-4809-811d-5fc94e141dd4","Type":"ContainerStarted","Data":"d0131aae0cb41e9c164bd273105c0aaa908cfe0682e90e9d454b1d148b26a5ab"} Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.133193 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.155316 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9rgf" event={"ID":"2f9658c9-2fb3-44c1-bb34-62a2e16360bb","Type":"ContainerStarted","Data":"0f2ae8dd9d5bb8ae41f96938378db82e98ffec4ed4783e12487b05071c50faa0"} Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.155362 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9rgf" event={"ID":"2f9658c9-2fb3-44c1-bb34-62a2e16360bb","Type":"ContainerStarted","Data":"0854dc9dc852cec7bffa0a79f6331a08faeddf4f61a5f2950908a89cfd526544"} Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.156948 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" podStartSLOduration=136.156923318 podStartE2EDuration="2m16.156923318s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:47.155042627 +0000 UTC m=+162.953514811" watchObservedRunningTime="2025-10-03 12:53:47.156923318 +0000 UTC m=+162.955395502" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.157259 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rb78z" podStartSLOduration=136.157253528 podStartE2EDuration="2m16.157253528s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:47.083702428 +0000 UTC m=+162.882174612" watchObservedRunningTime="2025-10-03 12:53:47.157253528 +0000 UTC m=+162.955725712" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.162622 4578 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-g5m4p container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" start-of-body= Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.162728 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" podUID="ac7d947f-4847-4809-811d-5fc94e141dd4" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.185562 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:47 crc kubenswrapper[4578]: E1003 12:53:47.185712 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:47.685693325 +0000 UTC m=+163.484165519 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.186052 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:47 crc kubenswrapper[4578]: E1003 12:53:47.187835 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:47.687824694 +0000 UTC m=+163.486296878 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.195398 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cxghf" event={"ID":"bc81f2fa-5ee7-4034-81d3-a11639762ac9","Type":"ContainerStarted","Data":"b7fc07878e6e422c4e8526e51122fcc940df0b14556e90498fff71d552447120"} Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.198740 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4" podStartSLOduration=136.198699625 podStartE2EDuration="2m16.198699625s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:47.195134229 +0000 UTC m=+162.993606423" watchObservedRunningTime="2025-10-03 12:53:47.198699625 +0000 UTC m=+162.997171809" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.228407 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" event={"ID":"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e","Type":"ContainerStarted","Data":"ad8012c352db1e465801c2720df816a0a606f653694a1696df01fa45a7494015"} Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.229524 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.230437 4578 generic.go:334] "Generic (PLEG): container finished" podID="bec2b13f-9bf4-4378-8a9c-56e61397071a" containerID="12d3409b652970b4921101fd7235ad7a4de846f69ee1aa15232a9945cd6fb199" exitCode=0 Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.230506 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" event={"ID":"bec2b13f-9bf4-4378-8a9c-56e61397071a","Type":"ContainerDied","Data":"12d3409b652970b4921101fd7235ad7a4de846f69ee1aa15232a9945cd6fb199"} Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.239764 4578 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-7br6x container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.7:6443/healthz\": dial tcp 10.217.0.7:6443: connect: connection refused" start-of-body= Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.239824 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" podUID="89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.7:6443/healthz\": dial tcp 10.217.0.7:6443: connect: connection refused" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.246815 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b" event={"ID":"5581f00e-5f39-454d-843d-9dbf6d2cbb0b","Type":"ContainerStarted","Data":"40b963f309694044275bec225a6a582161a7e0e9c5265e7a30930cfc924eefb7"} Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.259801 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-5bv45" event={"ID":"6c98467a-2f3b-4329-8f21-9492b087b5d8","Type":"ContainerStarted","Data":"94133ede5ae46a03ede26c3cdfae443c4e1e80c5c460cb075e84cd01fbe9480b"} Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.262717 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-rtlhn" event={"ID":"6528b192-d6c1-46e6-8048-58a76459651a","Type":"ContainerStarted","Data":"f6e667e4adb1ed62c57438cedd003f195e4e0d1d76e6d9fd04823a03cb79d50a"} Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.273764 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4qsl8" event={"ID":"959dff27-2ebe-417a-b5be-3e814dff3797","Type":"ContainerStarted","Data":"bb8c75463c0760c60e23f9e2bcfbb8a9c6a1981708ae66ffbdc5a90e1f10b4c0"} Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.275247 4578 patch_prober.go:28] interesting pod/downloads-7954f5f757-d2rzv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.275291 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-d2rzv" podUID="de78c746-e145-48ba-8601-a3ff2c6b2b62" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.287483 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:47 crc kubenswrapper[4578]: E1003 12:53:47.288135 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:47.788102066 +0000 UTC m=+163.586574270 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.288437 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:47 crc kubenswrapper[4578]: E1003 12:53:47.293814 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:47.79379642 +0000 UTC m=+163.592268674 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.316216 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" podStartSLOduration=136.316175831 podStartE2EDuration="2m16.316175831s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:47.308868895 +0000 UTC m=+163.107341079" watchObservedRunningTime="2025-10-03 12:53:47.316175831 +0000 UTC m=+163.114648015" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.317263 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7kdcr" podStartSLOduration=136.317255816 podStartE2EDuration="2m16.317255816s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:47.247058944 +0000 UTC m=+163.045531128" watchObservedRunningTime="2025-10-03 12:53:47.317255816 +0000 UTC m=+163.115728000" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.323671 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.390067 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:47 crc kubenswrapper[4578]: E1003 12:53:47.391060 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:47.891034183 +0000 UTC m=+163.689506367 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.406132 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9rgf" podStartSLOduration=136.406098439 podStartE2EDuration="2m16.406098439s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:47.357550784 +0000 UTC m=+163.156022968" watchObservedRunningTime="2025-10-03 12:53:47.406098439 +0000 UTC m=+163.204570623" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.406908 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz" podStartSLOduration=136.406901075 podStartE2EDuration="2m16.406901075s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:47.403849636 +0000 UTC m=+163.202321820" watchObservedRunningTime="2025-10-03 12:53:47.406901075 +0000 UTC m=+163.205373259" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.492251 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:47 crc kubenswrapper[4578]: E1003 12:53:47.492573 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:47.992561626 +0000 UTC m=+163.791033810 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.499563 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-5bv45" podStartSLOduration=136.499548731 podStartE2EDuration="2m16.499548731s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:47.498984813 +0000 UTC m=+163.297457007" watchObservedRunningTime="2025-10-03 12:53:47.499548731 +0000 UTC m=+163.298020915" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.603537 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:47 crc kubenswrapper[4578]: E1003 12:53:47.603820 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:48.10379012 +0000 UTC m=+163.902262314 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.604124 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:47 crc kubenswrapper[4578]: E1003 12:53:47.604411 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:48.1043968 +0000 UTC m=+163.902868984 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.706255 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:47 crc kubenswrapper[4578]: E1003 12:53:47.706703 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:48.206685067 +0000 UTC m=+164.005157261 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.729233 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4qsl8" podStartSLOduration=136.729211532 podStartE2EDuration="2m16.729211532s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:47.727820718 +0000 UTC m=+163.526292912" watchObservedRunningTime="2025-10-03 12:53:47.729211532 +0000 UTC m=+163.527683726" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.781742 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-rtlhn" podStartSLOduration=7.781728426 podStartE2EDuration="7.781728426s" podCreationTimestamp="2025-10-03 12:53:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:47.780442244 +0000 UTC m=+163.578914428" watchObservedRunningTime="2025-10-03 12:53:47.781728426 +0000 UTC m=+163.580200610" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.807285 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:47 crc kubenswrapper[4578]: E1003 12:53:47.807623 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:48.307611879 +0000 UTC m=+164.106084063 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.908071 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:47 crc kubenswrapper[4578]: E1003 12:53:47.908339 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:48.408325276 +0000 UTC m=+164.206797450 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.936406 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cxghf" podStartSLOduration=137.93638838 podStartE2EDuration="2m17.93638838s" podCreationTimestamp="2025-10-03 12:51:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:47.935158911 +0000 UTC m=+163.733631095" watchObservedRunningTime="2025-10-03 12:53:47.93638838 +0000 UTC m=+163.734860564" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.937582 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" podStartSLOduration=137.937576948 podStartE2EDuration="2m17.937576948s" podCreationTimestamp="2025-10-03 12:51:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:47.862031603 +0000 UTC m=+163.660503787" watchObservedRunningTime="2025-10-03 12:53:47.937576948 +0000 UTC m=+163.736049132" Oct 03 12:53:47 crc kubenswrapper[4578]: I1003 12:53:47.955949 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-bxmjt" Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.009922 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:48 crc kubenswrapper[4578]: E1003 12:53:48.010324 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:48.510307872 +0000 UTC m=+164.308780066 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.051696 4578 patch_prober.go:28] interesting pod/router-default-5444994796-q8kdg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 12:53:48 crc kubenswrapper[4578]: [-]has-synced failed: reason withheld Oct 03 12:53:48 crc kubenswrapper[4578]: [+]process-running ok Oct 03 12:53:48 crc kubenswrapper[4578]: healthz check failed Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.051814 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q8kdg" podUID="59bf594f-62b6-4a2f-a50d-30b5d8357133" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.111673 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:48 crc kubenswrapper[4578]: E1003 12:53:48.111893 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:48.611869025 +0000 UTC m=+164.410341209 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.112127 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:48 crc kubenswrapper[4578]: E1003 12:53:48.112447 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:48.612435684 +0000 UTC m=+164.410907868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.213078 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:48 crc kubenswrapper[4578]: E1003 12:53:48.213214 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:48.713196691 +0000 UTC m=+164.511668875 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.213340 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:48 crc kubenswrapper[4578]: E1003 12:53:48.213642 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:48.713622455 +0000 UTC m=+164.512094639 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.279859 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xctz5" event={"ID":"db43e12f-df81-44cf-9220-2f75d5ecc70b","Type":"ContainerStarted","Data":"cd0815b6f951415ac1068e2e8e328fcc57988e50bc8d12e3888d7125826055d1"} Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.282558 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-77kj8" event={"ID":"7d9d9160-dd25-4640-a84b-ee9558be3b88","Type":"ContainerStarted","Data":"be803f10fa99ad17fe7beb1ab5ff16ba150c9df7563828a8b308f11ae6706e41"} Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.284290 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4" event={"ID":"fdfd4313-d05e-4cbd-8520-3f929675c662","Type":"ContainerStarted","Data":"05ee2bd4e95415aa471a093fa842b223d3543a457c885f2d0816faaf04cf5fd6"} Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.286495 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-z7c9x" event={"ID":"86765ff0-85f5-462b-9141-0622f547152f","Type":"ContainerStarted","Data":"bc5bda99b4f7199724de78ac83223167c8d1267cf2434e9d3643d8ec7addeac1"} Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.286525 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-z7c9x" event={"ID":"86765ff0-85f5-462b-9141-0622f547152f","Type":"ContainerStarted","Data":"ec051363c33557517629d829214fed4dea523f051e25b29d0a3c1f65f9ea044d"} Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.286648 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-z7c9x" Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.287821 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-fgghg" event={"ID":"1c26320a-db2d-4b15-84ff-4b70590a946a","Type":"ContainerStarted","Data":"8e41afc03dc53db120dbbe535a4ba010c041d61d816f2e28e2a245c9ca664472"} Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.290239 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8htrd" event={"ID":"c48e4a1e-a25a-49d8-aa80-bd1e7425ae33","Type":"ContainerStarted","Data":"4bd7e6af50092ee3ca0f1ddb6c7c5d308af599004030a3978ad8c38a15f50706"} Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.290574 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8htrd" Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.292197 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" event={"ID":"bec2b13f-9bf4-4378-8a9c-56e61397071a","Type":"ContainerStarted","Data":"310e32fb61ea9c31fb511e0bb484acfb280ff7a51eb5f142c266410472640cbf"} Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.294544 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" event={"ID":"ad129323-bfad-47a1-84a5-7b51025f5d0d","Type":"ContainerStarted","Data":"c6118c1ea9c997885aea77f5c616e21d09a4e664fa984313f300b9af09077e33"} Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.294575 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" event={"ID":"ad129323-bfad-47a1-84a5-7b51025f5d0d","Type":"ContainerStarted","Data":"6f733da1cad8251b26486cc5cfb6f845ed25e17204fe616e3ba8788dd2602294"} Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.296733 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-5bv45" event={"ID":"6c98467a-2f3b-4329-8f21-9492b087b5d8","Type":"ContainerStarted","Data":"520aec33d980a001fd95d1bf3a737a095613b9f427e4510207e7b721a88128c3"} Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.298833 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-r9hlj" event={"ID":"f78b5fb4-0a61-4222-928f-2383bac97358","Type":"ContainerStarted","Data":"08f55de7e9fb56cd02955a3a23c68418680f71658838275708245067adcbf23b"} Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.300451 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" event={"ID":"ed3f40b1-ee95-427b-a6ea-0b412dba932c","Type":"ContainerStarted","Data":"5c50162488bbb6d491175f08575e7a3dce1264fc6ee9e05766faf6c79530087d"} Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.301556 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp" event={"ID":"b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02","Type":"ContainerStarted","Data":"29bdb747683b65c887f84f5dba2bf3b1a3bed3ed4af66c1389e5ae12cf73dc80"} Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.302675 4578 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-dd9jp container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:8443/healthz\": dial tcp 10.217.0.42:8443: connect: connection refused" start-of-body= Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.302721 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp" podUID="b0d5f336-b4f3-4aef-a99f-a5fe2b6a5c02" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.42:8443/healthz\": dial tcp 10.217.0.42:8443: connect: connection refused" Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.304162 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m68m5" event={"ID":"fcdc20ce-6b84-4c46-bec6-f1619fd6c2bc","Type":"ContainerStarted","Data":"b64e8c67ac6465f90ff3715b39d5e4cfaab7ede84bd311ea072eb76e9c738dd7"} Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.305809 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b" event={"ID":"5581f00e-5f39-454d-843d-9dbf6d2cbb0b","Type":"ContainerStarted","Data":"39ef58a4e20267c7db3c6d0887ed240ea065c367221680ca14cb2b7b7824203f"} Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.306723 4578 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-sqbkd container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.306767 4578 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-g5m4p container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" start-of-body= Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.306766 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" podUID="4ef05574-4924-4a73-b0ba-0e632c96410a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.306792 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" podUID="ac7d947f-4847-4809-811d-5fc94e141dd4" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.312344 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-l9fbz" Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.313733 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:48 crc kubenswrapper[4578]: E1003 12:53:48.313857 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:48.813841265 +0000 UTC m=+164.612313449 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.315289 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:48 crc kubenswrapper[4578]: E1003 12:53:48.315871 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:48.81586258 +0000 UTC m=+164.614334854 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.384819 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xctz5" podStartSLOduration=137.384797232 podStartE2EDuration="2m17.384797232s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:48.37389475 +0000 UTC m=+164.172366934" watchObservedRunningTime="2025-10-03 12:53:48.384797232 +0000 UTC m=+164.183269426" Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.417316 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:48 crc kubenswrapper[4578]: E1003 12:53:48.417511 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:48.917482595 +0000 UTC m=+164.715954789 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.419278 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:48 crc kubenswrapper[4578]: E1003 12:53:48.419355 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:48.919340575 +0000 UTC m=+164.717812849 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.460301 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-z7c9x" podStartSLOduration=8.460277254 podStartE2EDuration="8.460277254s" podCreationTimestamp="2025-10-03 12:53:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:48.419919923 +0000 UTC m=+164.218392117" watchObservedRunningTime="2025-10-03 12:53:48.460277254 +0000 UTC m=+164.258749438" Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.505409 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-r9hlj" podStartSLOduration=137.505388588 podStartE2EDuration="2m17.505388588s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:48.505220893 +0000 UTC m=+164.303693077" watchObservedRunningTime="2025-10-03 12:53:48.505388588 +0000 UTC m=+164.303860772" Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.546673 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:48 crc kubenswrapper[4578]: E1003 12:53:48.546987 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:49.046969488 +0000 UTC m=+164.845441672 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.614588 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.647219 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8htrd" podStartSLOduration=137.647201709 podStartE2EDuration="2m17.647201709s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:48.635626325 +0000 UTC m=+164.434098519" watchObservedRunningTime="2025-10-03 12:53:48.647201709 +0000 UTC m=+164.445673893" Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.648027 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:48 crc kubenswrapper[4578]: E1003 12:53:48.648899 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:49.148880532 +0000 UTC m=+164.947352726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.735715 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" podStartSLOduration=137.735696501 podStartE2EDuration="2m17.735696501s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:48.68914836 +0000 UTC m=+164.487620544" watchObservedRunningTime="2025-10-03 12:53:48.735696501 +0000 UTC m=+164.534168685" Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.749466 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:48 crc kubenswrapper[4578]: E1003 12:53:48.749685 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:49.249661031 +0000 UTC m=+165.048133215 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.749732 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:48 crc kubenswrapper[4578]: E1003 12:53:48.750055 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:49.250043693 +0000 UTC m=+165.048515877 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.839803 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m68m5" podStartSLOduration=137.839783886 podStartE2EDuration="2m17.839783886s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:48.739013328 +0000 UTC m=+164.537485512" watchObservedRunningTime="2025-10-03 12:53:48.839783886 +0000 UTC m=+164.638256070" Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.850261 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:48 crc kubenswrapper[4578]: E1003 12:53:48.850580 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:49.350566933 +0000 UTC m=+165.149039117 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.888879 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" podStartSLOduration=138.888851307 podStartE2EDuration="2m18.888851307s" podCreationTimestamp="2025-10-03 12:51:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:48.88772683 +0000 UTC m=+164.686199014" watchObservedRunningTime="2025-10-03 12:53:48.888851307 +0000 UTC m=+164.687323491" Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.889190 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-77kj8" podStartSLOduration=137.889184227 podStartE2EDuration="2m17.889184227s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:48.840710185 +0000 UTC m=+164.639182369" watchObservedRunningTime="2025-10-03 12:53:48.889184227 +0000 UTC m=+164.687656411" Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.941868 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2z8b" podStartSLOduration=137.941851425 podStartE2EDuration="2m17.941851425s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:48.939870341 +0000 UTC m=+164.738342535" watchObservedRunningTime="2025-10-03 12:53:48.941851425 +0000 UTC m=+164.740323609" Oct 03 12:53:48 crc kubenswrapper[4578]: I1003 12:53:48.952254 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:48 crc kubenswrapper[4578]: E1003 12:53:48.952553 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:49.452541729 +0000 UTC m=+165.251013913 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.048745 4578 patch_prober.go:28] interesting pod/router-default-5444994796-q8kdg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 12:53:49 crc kubenswrapper[4578]: [-]has-synced failed: reason withheld Oct 03 12:53:49 crc kubenswrapper[4578]: [+]process-running ok Oct 03 12:53:49 crc kubenswrapper[4578]: healthz check failed Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.048795 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q8kdg" podUID="59bf594f-62b6-4a2f-a50d-30b5d8357133" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.052769 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:49 crc kubenswrapper[4578]: E1003 12:53:49.052878 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:49.552859633 +0000 UTC m=+165.351331817 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.053175 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:49 crc kubenswrapper[4578]: E1003 12:53:49.053466 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:49.553456022 +0000 UTC m=+165.351928206 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.154026 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:49 crc kubenswrapper[4578]: E1003 12:53:49.154332 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:49.654302632 +0000 UTC m=+165.452774816 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.154623 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:49 crc kubenswrapper[4578]: E1003 12:53:49.154948 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:49.654935222 +0000 UTC m=+165.453407406 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.255679 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:49 crc kubenswrapper[4578]: E1003 12:53:49.256025 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:49.756010169 +0000 UTC m=+165.554482353 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.341409 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-dd9jp" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.357545 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:49 crc kubenswrapper[4578]: E1003 12:53:49.358914 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:49.858902726 +0000 UTC m=+165.657374910 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.458945 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:49 crc kubenswrapper[4578]: E1003 12:53:49.460131 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:49.960111057 +0000 UTC m=+165.758583261 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.560654 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:49 crc kubenswrapper[4578]: E1003 12:53:49.561122 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:50.061103592 +0000 UTC m=+165.859575846 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.622300 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ldfq4"] Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.623321 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ldfq4" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.636497 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.663655 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ldfq4"] Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.664538 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:49 crc kubenswrapper[4578]: E1003 12:53:49.665044 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:50.165027842 +0000 UTC m=+165.963500026 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.766183 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t287h\" (UniqueName: \"kubernetes.io/projected/b40b7869-c4f0-427d-b99c-fca7f0d8b937-kube-api-access-t287h\") pod \"community-operators-ldfq4\" (UID: \"b40b7869-c4f0-427d-b99c-fca7f0d8b937\") " pod="openshift-marketplace/community-operators-ldfq4" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.766255 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.766296 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b40b7869-c4f0-427d-b99c-fca7f0d8b937-utilities\") pod \"community-operators-ldfq4\" (UID: \"b40b7869-c4f0-427d-b99c-fca7f0d8b937\") " pod="openshift-marketplace/community-operators-ldfq4" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.766373 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b40b7869-c4f0-427d-b99c-fca7f0d8b937-catalog-content\") pod \"community-operators-ldfq4\" (UID: \"b40b7869-c4f0-427d-b99c-fca7f0d8b937\") " pod="openshift-marketplace/community-operators-ldfq4" Oct 03 12:53:49 crc kubenswrapper[4578]: E1003 12:53:49.766665 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:50.266647637 +0000 UTC m=+166.065119821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.790508 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hmk8t"] Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.791475 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hmk8t" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.795175 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.839669 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hmk8t"] Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.867140 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.867349 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b40b7869-c4f0-427d-b99c-fca7f0d8b937-catalog-content\") pod \"community-operators-ldfq4\" (UID: \"b40b7869-c4f0-427d-b99c-fca7f0d8b937\") " pod="openshift-marketplace/community-operators-ldfq4" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.867386 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/742a4571-0a86-4cb4-bb93-8b46219f6d49-catalog-content\") pod \"certified-operators-hmk8t\" (UID: \"742a4571-0a86-4cb4-bb93-8b46219f6d49\") " pod="openshift-marketplace/certified-operators-hmk8t" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.867404 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rccr\" (UniqueName: \"kubernetes.io/projected/742a4571-0a86-4cb4-bb93-8b46219f6d49-kube-api-access-7rccr\") pod \"certified-operators-hmk8t\" (UID: \"742a4571-0a86-4cb4-bb93-8b46219f6d49\") " pod="openshift-marketplace/certified-operators-hmk8t" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.867441 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/742a4571-0a86-4cb4-bb93-8b46219f6d49-utilities\") pod \"certified-operators-hmk8t\" (UID: \"742a4571-0a86-4cb4-bb93-8b46219f6d49\") " pod="openshift-marketplace/certified-operators-hmk8t" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.867467 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t287h\" (UniqueName: \"kubernetes.io/projected/b40b7869-c4f0-427d-b99c-fca7f0d8b937-kube-api-access-t287h\") pod \"community-operators-ldfq4\" (UID: \"b40b7869-c4f0-427d-b99c-fca7f0d8b937\") " pod="openshift-marketplace/community-operators-ldfq4" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.867514 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b40b7869-c4f0-427d-b99c-fca7f0d8b937-utilities\") pod \"community-operators-ldfq4\" (UID: \"b40b7869-c4f0-427d-b99c-fca7f0d8b937\") " pod="openshift-marketplace/community-operators-ldfq4" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.868080 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b40b7869-c4f0-427d-b99c-fca7f0d8b937-utilities\") pod \"community-operators-ldfq4\" (UID: \"b40b7869-c4f0-427d-b99c-fca7f0d8b937\") " pod="openshift-marketplace/community-operators-ldfq4" Oct 03 12:53:49 crc kubenswrapper[4578]: E1003 12:53:49.868170 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:50.368152739 +0000 UTC m=+166.166624923 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.868411 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b40b7869-c4f0-427d-b99c-fca7f0d8b937-catalog-content\") pod \"community-operators-ldfq4\" (UID: \"b40b7869-c4f0-427d-b99c-fca7f0d8b937\") " pod="openshift-marketplace/community-operators-ldfq4" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.935762 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t287h\" (UniqueName: \"kubernetes.io/projected/b40b7869-c4f0-427d-b99c-fca7f0d8b937-kube-api-access-t287h\") pod \"community-operators-ldfq4\" (UID: \"b40b7869-c4f0-427d-b99c-fca7f0d8b937\") " pod="openshift-marketplace/community-operators-ldfq4" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.938899 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ldfq4" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.969478 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/742a4571-0a86-4cb4-bb93-8b46219f6d49-catalog-content\") pod \"certified-operators-hmk8t\" (UID: \"742a4571-0a86-4cb4-bb93-8b46219f6d49\") " pod="openshift-marketplace/certified-operators-hmk8t" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.969773 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rccr\" (UniqueName: \"kubernetes.io/projected/742a4571-0a86-4cb4-bb93-8b46219f6d49-kube-api-access-7rccr\") pod \"certified-operators-hmk8t\" (UID: \"742a4571-0a86-4cb4-bb93-8b46219f6d49\") " pod="openshift-marketplace/certified-operators-hmk8t" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.969811 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/742a4571-0a86-4cb4-bb93-8b46219f6d49-utilities\") pod \"certified-operators-hmk8t\" (UID: \"742a4571-0a86-4cb4-bb93-8b46219f6d49\") " pod="openshift-marketplace/certified-operators-hmk8t" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.969843 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:49 crc kubenswrapper[4578]: E1003 12:53:49.970127 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:50.470115345 +0000 UTC m=+166.268587529 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.970602 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/742a4571-0a86-4cb4-bb93-8b46219f6d49-catalog-content\") pod \"certified-operators-hmk8t\" (UID: \"742a4571-0a86-4cb4-bb93-8b46219f6d49\") " pod="openshift-marketplace/certified-operators-hmk8t" Oct 03 12:53:49 crc kubenswrapper[4578]: I1003 12:53:49.971096 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/742a4571-0a86-4cb4-bb93-8b46219f6d49-utilities\") pod \"certified-operators-hmk8t\" (UID: \"742a4571-0a86-4cb4-bb93-8b46219f6d49\") " pod="openshift-marketplace/certified-operators-hmk8t" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.000896 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k4z8f"] Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.001796 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k4z8f" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.003069 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rccr\" (UniqueName: \"kubernetes.io/projected/742a4571-0a86-4cb4-bb93-8b46219f6d49-kube-api-access-7rccr\") pod \"certified-operators-hmk8t\" (UID: \"742a4571-0a86-4cb4-bb93-8b46219f6d49\") " pod="openshift-marketplace/certified-operators-hmk8t" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.055716 4578 patch_prober.go:28] interesting pod/router-default-5444994796-q8kdg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 12:53:50 crc kubenswrapper[4578]: [-]has-synced failed: reason withheld Oct 03 12:53:50 crc kubenswrapper[4578]: [+]process-running ok Oct 03 12:53:50 crc kubenswrapper[4578]: healthz check failed Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.055764 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q8kdg" podUID="59bf594f-62b6-4a2f-a50d-30b5d8357133" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.070894 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.071065 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55rzg\" (UniqueName: \"kubernetes.io/projected/445697f9-cf9b-44d0-b76a-c75e6910b758-kube-api-access-55rzg\") pod \"community-operators-k4z8f\" (UID: \"445697f9-cf9b-44d0-b76a-c75e6910b758\") " pod="openshift-marketplace/community-operators-k4z8f" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.071107 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/445697f9-cf9b-44d0-b76a-c75e6910b758-catalog-content\") pod \"community-operators-k4z8f\" (UID: \"445697f9-cf9b-44d0-b76a-c75e6910b758\") " pod="openshift-marketplace/community-operators-k4z8f" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.071160 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/445697f9-cf9b-44d0-b76a-c75e6910b758-utilities\") pod \"community-operators-k4z8f\" (UID: \"445697f9-cf9b-44d0-b76a-c75e6910b758\") " pod="openshift-marketplace/community-operators-k4z8f" Oct 03 12:53:50 crc kubenswrapper[4578]: E1003 12:53:50.071296 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:50.571264215 +0000 UTC m=+166.369736429 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.076949 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k4z8f"] Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.102569 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hmk8t" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.172109 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/445697f9-cf9b-44d0-b76a-c75e6910b758-catalog-content\") pod \"community-operators-k4z8f\" (UID: \"445697f9-cf9b-44d0-b76a-c75e6910b758\") " pod="openshift-marketplace/community-operators-k4z8f" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.172168 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.172189 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/445697f9-cf9b-44d0-b76a-c75e6910b758-utilities\") pod \"community-operators-k4z8f\" (UID: \"445697f9-cf9b-44d0-b76a-c75e6910b758\") " pod="openshift-marketplace/community-operators-k4z8f" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.172239 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55rzg\" (UniqueName: \"kubernetes.io/projected/445697f9-cf9b-44d0-b76a-c75e6910b758-kube-api-access-55rzg\") pod \"community-operators-k4z8f\" (UID: \"445697f9-cf9b-44d0-b76a-c75e6910b758\") " pod="openshift-marketplace/community-operators-k4z8f" Oct 03 12:53:50 crc kubenswrapper[4578]: E1003 12:53:50.173109 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:50.673093576 +0000 UTC m=+166.471565760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.173270 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/445697f9-cf9b-44d0-b76a-c75e6910b758-catalog-content\") pod \"community-operators-k4z8f\" (UID: \"445697f9-cf9b-44d0-b76a-c75e6910b758\") " pod="openshift-marketplace/community-operators-k4z8f" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.173565 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/445697f9-cf9b-44d0-b76a-c75e6910b758-utilities\") pod \"community-operators-k4z8f\" (UID: \"445697f9-cf9b-44d0-b76a-c75e6910b758\") " pod="openshift-marketplace/community-operators-k4z8f" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.173985 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g5m4p" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.238778 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-94bd2"] Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.239682 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-94bd2" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.262162 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-94bd2"] Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.273687 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:50 crc kubenswrapper[4578]: E1003 12:53:50.273982 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:50.773967277 +0000 UTC m=+166.572439461 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.279241 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55rzg\" (UniqueName: \"kubernetes.io/projected/445697f9-cf9b-44d0-b76a-c75e6910b758-kube-api-access-55rzg\") pod \"community-operators-k4z8f\" (UID: \"445697f9-cf9b-44d0-b76a-c75e6910b758\") " pod="openshift-marketplace/community-operators-k4z8f" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.333366 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k4z8f" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.378994 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2244329a-bbb1-4c5a-a170-55106dcb21c2-catalog-content\") pod \"certified-operators-94bd2\" (UID: \"2244329a-bbb1-4c5a-a170-55106dcb21c2\") " pod="openshift-marketplace/certified-operators-94bd2" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.379044 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2244329a-bbb1-4c5a-a170-55106dcb21c2-utilities\") pod \"certified-operators-94bd2\" (UID: \"2244329a-bbb1-4c5a-a170-55106dcb21c2\") " pod="openshift-marketplace/certified-operators-94bd2" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.379083 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.379197 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws6gc\" (UniqueName: \"kubernetes.io/projected/2244329a-bbb1-4c5a-a170-55106dcb21c2-kube-api-access-ws6gc\") pod \"certified-operators-94bd2\" (UID: \"2244329a-bbb1-4c5a-a170-55106dcb21c2\") " pod="openshift-marketplace/certified-operators-94bd2" Oct 03 12:53:50 crc kubenswrapper[4578]: E1003 12:53:50.380710 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:50.880695628 +0000 UTC m=+166.679167812 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.481230 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.481372 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2244329a-bbb1-4c5a-a170-55106dcb21c2-catalog-content\") pod \"certified-operators-94bd2\" (UID: \"2244329a-bbb1-4c5a-a170-55106dcb21c2\") " pod="openshift-marketplace/certified-operators-94bd2" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.481402 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2244329a-bbb1-4c5a-a170-55106dcb21c2-utilities\") pod \"certified-operators-94bd2\" (UID: \"2244329a-bbb1-4c5a-a170-55106dcb21c2\") " pod="openshift-marketplace/certified-operators-94bd2" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.481453 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws6gc\" (UniqueName: \"kubernetes.io/projected/2244329a-bbb1-4c5a-a170-55106dcb21c2-kube-api-access-ws6gc\") pod \"certified-operators-94bd2\" (UID: \"2244329a-bbb1-4c5a-a170-55106dcb21c2\") " pod="openshift-marketplace/certified-operators-94bd2" Oct 03 12:53:50 crc kubenswrapper[4578]: E1003 12:53:50.481873 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:50.981855017 +0000 UTC m=+166.780327201 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.482223 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2244329a-bbb1-4c5a-a170-55106dcb21c2-catalog-content\") pod \"certified-operators-94bd2\" (UID: \"2244329a-bbb1-4c5a-a170-55106dcb21c2\") " pod="openshift-marketplace/certified-operators-94bd2" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.485959 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2244329a-bbb1-4c5a-a170-55106dcb21c2-utilities\") pod \"certified-operators-94bd2\" (UID: \"2244329a-bbb1-4c5a-a170-55106dcb21c2\") " pod="openshift-marketplace/certified-operators-94bd2" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.566913 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws6gc\" (UniqueName: \"kubernetes.io/projected/2244329a-bbb1-4c5a-a170-55106dcb21c2-kube-api-access-ws6gc\") pod \"certified-operators-94bd2\" (UID: \"2244329a-bbb1-4c5a-a170-55106dcb21c2\") " pod="openshift-marketplace/certified-operators-94bd2" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.586292 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:50 crc kubenswrapper[4578]: E1003 12:53:50.586608 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:51.086598063 +0000 UTC m=+166.885070247 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.688454 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:50 crc kubenswrapper[4578]: E1003 12:53:50.688848 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:51.188830258 +0000 UTC m=+166.987302452 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.789885 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:50 crc kubenswrapper[4578]: E1003 12:53:50.790178 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:51.290167734 +0000 UTC m=+167.088639908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.857236 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-94bd2" Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.879405 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ldfq4"] Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.890706 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:50 crc kubenswrapper[4578]: E1003 12:53:50.890931 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:51.390895301 +0000 UTC m=+167.189367485 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.891184 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:50 crc kubenswrapper[4578]: E1003 12:53:50.891581 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:51.391570842 +0000 UTC m=+167.190043026 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:50 crc kubenswrapper[4578]: I1003 12:53:50.992794 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:50 crc kubenswrapper[4578]: E1003 12:53:50.993343 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:51.493328132 +0000 UTC m=+167.291800316 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.023724 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hmk8t"] Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.073869 4578 patch_prober.go:28] interesting pod/router-default-5444994796-q8kdg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 12:53:51 crc kubenswrapper[4578]: [-]has-synced failed: reason withheld Oct 03 12:53:51 crc kubenswrapper[4578]: [+]process-running ok Oct 03 12:53:51 crc kubenswrapper[4578]: healthz check failed Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.074170 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q8kdg" podUID="59bf594f-62b6-4a2f-a50d-30b5d8357133" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.096250 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:51 crc kubenswrapper[4578]: E1003 12:53:51.096545 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:51.596530458 +0000 UTC m=+167.395002642 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.197326 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:51 crc kubenswrapper[4578]: E1003 12:53:51.197521 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:51.697502973 +0000 UTC m=+167.495975157 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.197615 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:51 crc kubenswrapper[4578]: E1003 12:53:51.197999 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:51.697977848 +0000 UTC m=+167.496450032 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.298481 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:51 crc kubenswrapper[4578]: E1003 12:53:51.298834 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:51.798820447 +0000 UTC m=+167.597292631 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.315530 4578 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-8htrd container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.315587 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8htrd" podUID="c48e4a1e-a25a-49d8-aa80-bd1e7425ae33" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.327393 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hmk8t" event={"ID":"742a4571-0a86-4cb4-bb93-8b46219f6d49","Type":"ContainerStarted","Data":"84225a706ab101fc2aa255d82534aea42e72fc84811bc366ef8b2ff78c78372c"} Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.330330 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ldfq4" event={"ID":"b40b7869-c4f0-427d-b99c-fca7f0d8b937","Type":"ContainerStarted","Data":"2fc756e984b124eb0a6d4c5d26df07dda09d9cbb1668492b67d8fb1fc589b67a"} Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.342436 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" event={"ID":"ed3f40b1-ee95-427b-a6ea-0b412dba932c","Type":"ContainerStarted","Data":"eda836a09e242d702f4255a6fdf7369d12ea8979af654e43a78ef4bcb2051510"} Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.401541 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:51 crc kubenswrapper[4578]: E1003 12:53:51.401909 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:51.901896509 +0000 UTC m=+167.700368693 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.404178 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k4z8f"] Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.505337 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:51 crc kubenswrapper[4578]: E1003 12:53:51.505969 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:52.005953994 +0000 UTC m=+167.804426178 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.607764 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:51 crc kubenswrapper[4578]: E1003 12:53:51.608057 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:52.108045424 +0000 UTC m=+167.906517608 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.635340 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-94bd2"] Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.709026 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:51 crc kubenswrapper[4578]: E1003 12:53:51.709392 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:52.209373959 +0000 UTC m=+168.007846143 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.713371 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8htrd" Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.786036 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w5hjv"] Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.789986 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w5hjv" Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.807102 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.810875 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:51 crc kubenswrapper[4578]: E1003 12:53:51.811143 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:52.311131419 +0000 UTC m=+168.109603603 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.911299 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5hjv"] Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.912975 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.913155 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75b24dc1-77ec-4b54-9f85-2443e273eadf-catalog-content\") pod \"redhat-marketplace-w5hjv\" (UID: \"75b24dc1-77ec-4b54-9f85-2443e273eadf\") " pod="openshift-marketplace/redhat-marketplace-w5hjv" Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.913217 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75b24dc1-77ec-4b54-9f85-2443e273eadf-utilities\") pod \"redhat-marketplace-w5hjv\" (UID: \"75b24dc1-77ec-4b54-9f85-2443e273eadf\") " pod="openshift-marketplace/redhat-marketplace-w5hjv" Oct 03 12:53:51 crc kubenswrapper[4578]: I1003 12:53:51.913260 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5bqk\" (UniqueName: \"kubernetes.io/projected/75b24dc1-77ec-4b54-9f85-2443e273eadf-kube-api-access-s5bqk\") pod \"redhat-marketplace-w5hjv\" (UID: \"75b24dc1-77ec-4b54-9f85-2443e273eadf\") " pod="openshift-marketplace/redhat-marketplace-w5hjv" Oct 03 12:53:51 crc kubenswrapper[4578]: E1003 12:53:51.913353 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:52.413340113 +0000 UTC m=+168.211812297 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.016447 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75b24dc1-77ec-4b54-9f85-2443e273eadf-utilities\") pod \"redhat-marketplace-w5hjv\" (UID: \"75b24dc1-77ec-4b54-9f85-2443e273eadf\") " pod="openshift-marketplace/redhat-marketplace-w5hjv" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.016532 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5bqk\" (UniqueName: \"kubernetes.io/projected/75b24dc1-77ec-4b54-9f85-2443e273eadf-kube-api-access-s5bqk\") pod \"redhat-marketplace-w5hjv\" (UID: \"75b24dc1-77ec-4b54-9f85-2443e273eadf\") " pod="openshift-marketplace/redhat-marketplace-w5hjv" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.016567 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75b24dc1-77ec-4b54-9f85-2443e273eadf-catalog-content\") pod \"redhat-marketplace-w5hjv\" (UID: \"75b24dc1-77ec-4b54-9f85-2443e273eadf\") " pod="openshift-marketplace/redhat-marketplace-w5hjv" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.016603 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.016945 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75b24dc1-77ec-4b54-9f85-2443e273eadf-utilities\") pod \"redhat-marketplace-w5hjv\" (UID: \"75b24dc1-77ec-4b54-9f85-2443e273eadf\") " pod="openshift-marketplace/redhat-marketplace-w5hjv" Oct 03 12:53:52 crc kubenswrapper[4578]: E1003 12:53:52.016962 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:52.516947642 +0000 UTC m=+168.315419826 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.017186 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75b24dc1-77ec-4b54-9f85-2443e273eadf-catalog-content\") pod \"redhat-marketplace-w5hjv\" (UID: \"75b24dc1-77ec-4b54-9f85-2443e273eadf\") " pod="openshift-marketplace/redhat-marketplace-w5hjv" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.054793 4578 patch_prober.go:28] interesting pod/router-default-5444994796-q8kdg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 12:53:52 crc kubenswrapper[4578]: [-]has-synced failed: reason withheld Oct 03 12:53:52 crc kubenswrapper[4578]: [+]process-running ok Oct 03 12:53:52 crc kubenswrapper[4578]: healthz check failed Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.054856 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q8kdg" podUID="59bf594f-62b6-4a2f-a50d-30b5d8357133" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.096295 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5bqk\" (UniqueName: \"kubernetes.io/projected/75b24dc1-77ec-4b54-9f85-2443e273eadf-kube-api-access-s5bqk\") pod \"redhat-marketplace-w5hjv\" (UID: \"75b24dc1-77ec-4b54-9f85-2443e273eadf\") " pod="openshift-marketplace/redhat-marketplace-w5hjv" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.107986 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w5hjv" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.119308 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:52 crc kubenswrapper[4578]: E1003 12:53:52.119589 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:52.61957565 +0000 UTC m=+168.418047824 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.168247 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qdvvk"] Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.169159 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qdvvk" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.186390 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qdvvk"] Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.220840 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc33b180-1050-403c-81b7-00ffdf579cd7-catalog-content\") pod \"redhat-marketplace-qdvvk\" (UID: \"fc33b180-1050-403c-81b7-00ffdf579cd7\") " pod="openshift-marketplace/redhat-marketplace-qdvvk" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.220906 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmvw5\" (UniqueName: \"kubernetes.io/projected/fc33b180-1050-403c-81b7-00ffdf579cd7-kube-api-access-xmvw5\") pod \"redhat-marketplace-qdvvk\" (UID: \"fc33b180-1050-403c-81b7-00ffdf579cd7\") " pod="openshift-marketplace/redhat-marketplace-qdvvk" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.220944 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.220981 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc33b180-1050-403c-81b7-00ffdf579cd7-utilities\") pod \"redhat-marketplace-qdvvk\" (UID: \"fc33b180-1050-403c-81b7-00ffdf579cd7\") " pod="openshift-marketplace/redhat-marketplace-qdvvk" Oct 03 12:53:52 crc kubenswrapper[4578]: E1003 12:53:52.221261 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:52.721249697 +0000 UTC m=+168.519721881 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.226308 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.227438 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.243755 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.243805 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.245024 4578 patch_prober.go:28] interesting pod/console-f9d7485db-ns26l container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.245074 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-ns26l" podUID="7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.263427 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.292618 4578 patch_prober.go:28] interesting pod/downloads-7954f5f757-d2rzv container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.292946 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-d2rzv" podUID="de78c746-e145-48ba-8601-a3ff2c6b2b62" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.292665 4578 patch_prober.go:28] interesting pod/downloads-7954f5f757-d2rzv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.293134 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-d2rzv" podUID="de78c746-e145-48ba-8601-a3ff2c6b2b62" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.322264 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.322553 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc33b180-1050-403c-81b7-00ffdf579cd7-utilities\") pod \"redhat-marketplace-qdvvk\" (UID: \"fc33b180-1050-403c-81b7-00ffdf579cd7\") " pod="openshift-marketplace/redhat-marketplace-qdvvk" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.323069 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc33b180-1050-403c-81b7-00ffdf579cd7-utilities\") pod \"redhat-marketplace-qdvvk\" (UID: \"fc33b180-1050-403c-81b7-00ffdf579cd7\") " pod="openshift-marketplace/redhat-marketplace-qdvvk" Oct 03 12:53:52 crc kubenswrapper[4578]: E1003 12:53:52.323173 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:52.823156011 +0000 UTC m=+168.621628185 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.323939 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc33b180-1050-403c-81b7-00ffdf579cd7-catalog-content\") pod \"redhat-marketplace-qdvvk\" (UID: \"fc33b180-1050-403c-81b7-00ffdf579cd7\") " pod="openshift-marketplace/redhat-marketplace-qdvvk" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.324018 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmvw5\" (UniqueName: \"kubernetes.io/projected/fc33b180-1050-403c-81b7-00ffdf579cd7-kube-api-access-xmvw5\") pod \"redhat-marketplace-qdvvk\" (UID: \"fc33b180-1050-403c-81b7-00ffdf579cd7\") " pod="openshift-marketplace/redhat-marketplace-qdvvk" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.324668 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc33b180-1050-403c-81b7-00ffdf579cd7-catalog-content\") pod \"redhat-marketplace-qdvvk\" (UID: \"fc33b180-1050-403c-81b7-00ffdf579cd7\") " pod="openshift-marketplace/redhat-marketplace-qdvvk" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.354451 4578 generic.go:334] "Generic (PLEG): container finished" podID="742a4571-0a86-4cb4-bb93-8b46219f6d49" containerID="5a8cdf1cca6e3eef1b5bafa3c1518bf4e4b7a593437be92d88c7fa1df46e83b7" exitCode=0 Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.354868 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hmk8t" event={"ID":"742a4571-0a86-4cb4-bb93-8b46219f6d49","Type":"ContainerDied","Data":"5a8cdf1cca6e3eef1b5bafa3c1518bf4e4b7a593437be92d88c7fa1df46e83b7"} Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.356336 4578 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.366473 4578 generic.go:334] "Generic (PLEG): container finished" podID="2244329a-bbb1-4c5a-a170-55106dcb21c2" containerID="6b5f2a5de88e4dd4a49eaa028f75af62c13d0a0fd88c6305b603f5cc8674e353" exitCode=0 Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.366569 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-94bd2" event={"ID":"2244329a-bbb1-4c5a-a170-55106dcb21c2","Type":"ContainerDied","Data":"6b5f2a5de88e4dd4a49eaa028f75af62c13d0a0fd88c6305b603f5cc8674e353"} Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.366594 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-94bd2" event={"ID":"2244329a-bbb1-4c5a-a170-55106dcb21c2","Type":"ContainerStarted","Data":"caee4924e18253ed96aae28aa62a3018f45ed38c5126987554a77a937ad688b3"} Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.379438 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmvw5\" (UniqueName: \"kubernetes.io/projected/fc33b180-1050-403c-81b7-00ffdf579cd7-kube-api-access-xmvw5\") pod \"redhat-marketplace-qdvvk\" (UID: \"fc33b180-1050-403c-81b7-00ffdf579cd7\") " pod="openshift-marketplace/redhat-marketplace-qdvvk" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.390106 4578 generic.go:334] "Generic (PLEG): container finished" podID="b40b7869-c4f0-427d-b99c-fca7f0d8b937" containerID="5d83e6a2765fabe8666fefde1a7eca38a5cbad1ebfb98e191920c7b5afa511e0" exitCode=0 Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.390317 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ldfq4" event={"ID":"b40b7869-c4f0-427d-b99c-fca7f0d8b937","Type":"ContainerDied","Data":"5d83e6a2765fabe8666fefde1a7eca38a5cbad1ebfb98e191920c7b5afa511e0"} Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.393215 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.393239 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.426725 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:52 crc kubenswrapper[4578]: E1003 12:53:52.429745 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:52.929719265 +0000 UTC m=+168.728191449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.460086 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" event={"ID":"ed3f40b1-ee95-427b-a6ea-0b412dba932c","Type":"ContainerStarted","Data":"d9e70cfc66f66e30c51997868acfaa8ba4a011e3e953fab80e2620d70aa4daa4"} Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.504577 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qdvvk" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.530612 4578 generic.go:334] "Generic (PLEG): container finished" podID="445697f9-cf9b-44d0-b76a-c75e6910b758" containerID="5cb73ff0eae1a1350588654698e7e1e42941c1bfac0342e201d4095cb5622199" exitCode=0 Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.531470 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4z8f" event={"ID":"445697f9-cf9b-44d0-b76a-c75e6910b758","Type":"ContainerDied","Data":"5cb73ff0eae1a1350588654698e7e1e42941c1bfac0342e201d4095cb5622199"} Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.531498 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4z8f" event={"ID":"445697f9-cf9b-44d0-b76a-c75e6910b758","Type":"ContainerStarted","Data":"431146cd9bcbc4268b4097e1cc64cbc660db407c1e4c3f9cb166bb02b223ac38"} Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.532442 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:52 crc kubenswrapper[4578]: E1003 12:53:52.532745 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:53.032731356 +0000 UTC m=+168.831203540 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.548817 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvkjq" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.634524 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:52 crc kubenswrapper[4578]: E1003 12:53:52.635150 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:53.135135985 +0000 UTC m=+168.933608169 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.735722 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:52 crc kubenswrapper[4578]: E1003 12:53:52.735856 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:53.235841772 +0000 UTC m=+169.034313956 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.736017 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:52 crc kubenswrapper[4578]: E1003 12:53:52.736279 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:53.236270435 +0000 UTC m=+169.034742619 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.742265 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.840103 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:52 crc kubenswrapper[4578]: E1003 12:53:52.841132 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:53.341117924 +0000 UTC m=+169.139590108 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.881964 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n6875"] Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.883014 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n6875" Oct 03 12:53:52 crc kubenswrapper[4578]: W1003 12:53:52.896447 4578 reflector.go:561] object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh": failed to list *v1.Secret: secrets "redhat-operators-dockercfg-ct8rh" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Oct 03 12:53:52 crc kubenswrapper[4578]: E1003 12:53:52.896488 4578 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"redhat-operators-dockercfg-ct8rh\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"redhat-operators-dockercfg-ct8rh\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.952688 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n6875"] Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.952898 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.952953 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfh44\" (UniqueName: \"kubernetes.io/projected/beb5b22e-cc9e-453b-bac2-88f7a5d30c29-kube-api-access-rfh44\") pod \"redhat-operators-n6875\" (UID: \"beb5b22e-cc9e-453b-bac2-88f7a5d30c29\") " pod="openshift-marketplace/redhat-operators-n6875" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.952989 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beb5b22e-cc9e-453b-bac2-88f7a5d30c29-utilities\") pod \"redhat-operators-n6875\" (UID: \"beb5b22e-cc9e-453b-bac2-88f7a5d30c29\") " pod="openshift-marketplace/redhat-operators-n6875" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.953007 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beb5b22e-cc9e-453b-bac2-88f7a5d30c29-catalog-content\") pod \"redhat-operators-n6875\" (UID: \"beb5b22e-cc9e-453b-bac2-88f7a5d30c29\") " pod="openshift-marketplace/redhat-operators-n6875" Oct 03 12:53:52 crc kubenswrapper[4578]: E1003 12:53:52.953301 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:53.45328742 +0000 UTC m=+169.251759604 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.954890 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.955506 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.977098 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 03 12:53:52 crc kubenswrapper[4578]: I1003 12:53:52.977546 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.019582 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5hjv"] Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.046516 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.055033 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.055228 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30c6876c-90e4-4670-999d-64556e223478-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"30c6876c-90e4-4670-999d-64556e223478\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.055271 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfh44\" (UniqueName: \"kubernetes.io/projected/beb5b22e-cc9e-453b-bac2-88f7a5d30c29-kube-api-access-rfh44\") pod \"redhat-operators-n6875\" (UID: \"beb5b22e-cc9e-453b-bac2-88f7a5d30c29\") " pod="openshift-marketplace/redhat-operators-n6875" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.055321 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beb5b22e-cc9e-453b-bac2-88f7a5d30c29-utilities\") pod \"redhat-operators-n6875\" (UID: \"beb5b22e-cc9e-453b-bac2-88f7a5d30c29\") " pod="openshift-marketplace/redhat-operators-n6875" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.055354 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beb5b22e-cc9e-453b-bac2-88f7a5d30c29-catalog-content\") pod \"redhat-operators-n6875\" (UID: \"beb5b22e-cc9e-453b-bac2-88f7a5d30c29\") " pod="openshift-marketplace/redhat-operators-n6875" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.055381 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/30c6876c-90e4-4670-999d-64556e223478-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"30c6876c-90e4-4670-999d-64556e223478\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 12:53:53 crc kubenswrapper[4578]: E1003 12:53:53.055483 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:53.555469373 +0000 UTC m=+169.353941557 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.056073 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beb5b22e-cc9e-453b-bac2-88f7a5d30c29-utilities\") pod \"redhat-operators-n6875\" (UID: \"beb5b22e-cc9e-453b-bac2-88f7a5d30c29\") " pod="openshift-marketplace/redhat-operators-n6875" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.056289 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beb5b22e-cc9e-453b-bac2-88f7a5d30c29-catalog-content\") pod \"redhat-operators-n6875\" (UID: \"beb5b22e-cc9e-453b-bac2-88f7a5d30c29\") " pod="openshift-marketplace/redhat-operators-n6875" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.057984 4578 patch_prober.go:28] interesting pod/router-default-5444994796-q8kdg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 12:53:53 crc kubenswrapper[4578]: [-]has-synced failed: reason withheld Oct 03 12:53:53 crc kubenswrapper[4578]: [+]process-running ok Oct 03 12:53:53 crc kubenswrapper[4578]: healthz check failed Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.058049 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q8kdg" podUID="59bf594f-62b6-4a2f-a50d-30b5d8357133" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.069932 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.138599 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfh44\" (UniqueName: \"kubernetes.io/projected/beb5b22e-cc9e-453b-bac2-88f7a5d30c29-kube-api-access-rfh44\") pod \"redhat-operators-n6875\" (UID: \"beb5b22e-cc9e-453b-bac2-88f7a5d30c29\") " pod="openshift-marketplace/redhat-operators-n6875" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.157355 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30c6876c-90e4-4670-999d-64556e223478-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"30c6876c-90e4-4670-999d-64556e223478\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.157429 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/30c6876c-90e4-4670-999d-64556e223478-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"30c6876c-90e4-4670-999d-64556e223478\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.157482 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:53 crc kubenswrapper[4578]: E1003 12:53:53.158922 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:53.658911347 +0000 UTC m=+169.457383531 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.158941 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/30c6876c-90e4-4670-999d-64556e223478-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"30c6876c-90e4-4670-999d-64556e223478\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.215233 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2nzxc"] Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.216211 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2nzxc" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.242879 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2nzxc"] Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.244105 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30c6876c-90e4-4670-999d-64556e223478-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"30c6876c-90e4-4670-999d-64556e223478\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.261813 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.262072 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4754126b-ff64-44ce-829d-6f8307eed3a7-catalog-content\") pod \"redhat-operators-2nzxc\" (UID: \"4754126b-ff64-44ce-829d-6f8307eed3a7\") " pod="openshift-marketplace/redhat-operators-2nzxc" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.262104 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr5cg\" (UniqueName: \"kubernetes.io/projected/4754126b-ff64-44ce-829d-6f8307eed3a7-kube-api-access-pr5cg\") pod \"redhat-operators-2nzxc\" (UID: \"4754126b-ff64-44ce-829d-6f8307eed3a7\") " pod="openshift-marketplace/redhat-operators-2nzxc" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.262128 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4754126b-ff64-44ce-829d-6f8307eed3a7-utilities\") pod \"redhat-operators-2nzxc\" (UID: \"4754126b-ff64-44ce-829d-6f8307eed3a7\") " pod="openshift-marketplace/redhat-operators-2nzxc" Oct 03 12:53:53 crc kubenswrapper[4578]: E1003 12:53:53.262222 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:53.762208616 +0000 UTC m=+169.560680800 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.312791 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.356971 4578 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.365028 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4754126b-ff64-44ce-829d-6f8307eed3a7-catalog-content\") pod \"redhat-operators-2nzxc\" (UID: \"4754126b-ff64-44ce-829d-6f8307eed3a7\") " pod="openshift-marketplace/redhat-operators-2nzxc" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.365066 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr5cg\" (UniqueName: \"kubernetes.io/projected/4754126b-ff64-44ce-829d-6f8307eed3a7-kube-api-access-pr5cg\") pod \"redhat-operators-2nzxc\" (UID: \"4754126b-ff64-44ce-829d-6f8307eed3a7\") " pod="openshift-marketplace/redhat-operators-2nzxc" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.365097 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4754126b-ff64-44ce-829d-6f8307eed3a7-utilities\") pod \"redhat-operators-2nzxc\" (UID: \"4754126b-ff64-44ce-829d-6f8307eed3a7\") " pod="openshift-marketplace/redhat-operators-2nzxc" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.365152 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.365195 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs\") pod \"network-metrics-daemon-q8986\" (UID: \"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\") " pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.368326 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4754126b-ff64-44ce-829d-6f8307eed3a7-catalog-content\") pod \"redhat-operators-2nzxc\" (UID: \"4754126b-ff64-44ce-829d-6f8307eed3a7\") " pod="openshift-marketplace/redhat-operators-2nzxc" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.368754 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4754126b-ff64-44ce-829d-6f8307eed3a7-utilities\") pod \"redhat-operators-2nzxc\" (UID: \"4754126b-ff64-44ce-829d-6f8307eed3a7\") " pod="openshift-marketplace/redhat-operators-2nzxc" Oct 03 12:53:53 crc kubenswrapper[4578]: E1003 12:53:53.369062 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:53.869047039 +0000 UTC m=+169.667519223 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.375436 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2971c7bb-e0f4-44a8-8f52-9ea8ba199772-metrics-certs\") pod \"network-metrics-daemon-q8986\" (UID: \"2971c7bb-e0f4-44a8-8f52-9ea8ba199772\") " pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.400452 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr5cg\" (UniqueName: \"kubernetes.io/projected/4754126b-ff64-44ce-829d-6f8307eed3a7-kube-api-access-pr5cg\") pod \"redhat-operators-2nzxc\" (UID: \"4754126b-ff64-44ce-829d-6f8307eed3a7\") " pod="openshift-marketplace/redhat-operators-2nzxc" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.423746 4578 patch_prober.go:28] interesting pod/apiserver-76f77b778f-wmkdd container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 03 12:53:53 crc kubenswrapper[4578]: [+]log ok Oct 03 12:53:53 crc kubenswrapper[4578]: [+]etcd ok Oct 03 12:53:53 crc kubenswrapper[4578]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 03 12:53:53 crc kubenswrapper[4578]: [+]poststarthook/generic-apiserver-start-informers ok Oct 03 12:53:53 crc kubenswrapper[4578]: [+]poststarthook/max-in-flight-filter ok Oct 03 12:53:53 crc kubenswrapper[4578]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 03 12:53:53 crc kubenswrapper[4578]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 03 12:53:53 crc kubenswrapper[4578]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 03 12:53:53 crc kubenswrapper[4578]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 03 12:53:53 crc kubenswrapper[4578]: [+]poststarthook/project.openshift.io-projectcache ok Oct 03 12:53:53 crc kubenswrapper[4578]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 03 12:53:53 crc kubenswrapper[4578]: [+]poststarthook/openshift.io-startinformers ok Oct 03 12:53:53 crc kubenswrapper[4578]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 03 12:53:53 crc kubenswrapper[4578]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 03 12:53:53 crc kubenswrapper[4578]: livez check failed Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.423806 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" podUID="ad129323-bfad-47a1-84a5-7b51025f5d0d" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.466257 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:53 crc kubenswrapper[4578]: E1003 12:53:53.466382 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:53.966357675 +0000 UTC m=+169.764829849 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.466453 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:53 crc kubenswrapper[4578]: E1003 12:53:53.466765 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:53.966757458 +0000 UTC m=+169.765229642 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.472758 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qdvvk"] Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.538006 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-q8986" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.567118 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:53 crc kubenswrapper[4578]: E1003 12:53:53.567370 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:54.067354871 +0000 UTC m=+169.865827055 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.578502 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdvvk" event={"ID":"fc33b180-1050-403c-81b7-00ffdf579cd7","Type":"ContainerStarted","Data":"a85ac3e1addae15d336b28b97794e316043c90bc5df28586a634f67ab69d9253"} Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.583265 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5hjv" event={"ID":"75b24dc1-77ec-4b54-9f85-2443e273eadf","Type":"ContainerStarted","Data":"bbf91797952d16a949c6632514a820d7c4fd06aeb5108932c7905c80d8358b78"} Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.583302 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5hjv" event={"ID":"75b24dc1-77ec-4b54-9f85-2443e273eadf","Type":"ContainerStarted","Data":"cc9b40b91a21e293024f14404d5388efaa244cbd41cf58892f488cbdebf1a7d2"} Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.629576 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" event={"ID":"ed3f40b1-ee95-427b-a6ea-0b412dba932c","Type":"ContainerStarted","Data":"67b3ad4e96e65cc99dfbdae88b8779d0473f7b392ccb687c2e02e0623553c2a4"} Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.665715 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-wc8cd" podStartSLOduration=13.665696 podStartE2EDuration="13.665696s" podCreationTimestamp="2025-10-03 12:53:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:53.664614295 +0000 UTC m=+169.463086479" watchObservedRunningTime="2025-10-03 12:53:53.665696 +0000 UTC m=+169.464168184" Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.672737 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:53 crc kubenswrapper[4578]: E1003 12:53:53.676693 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:54.176679614 +0000 UTC m=+169.975151798 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.779025 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:53 crc kubenswrapper[4578]: E1003 12:53:53.779344 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:54.279323873 +0000 UTC m=+170.077796057 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.779425 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:53 crc kubenswrapper[4578]: E1003 12:53:53.779833 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:54.279822099 +0000 UTC m=+170.078294283 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.885288 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:53 crc kubenswrapper[4578]: E1003 12:53:53.885608 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:54.385594117 +0000 UTC m=+170.184066301 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.969978 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 03 12:53:53 crc kubenswrapper[4578]: I1003 12:53:53.986772 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:53 crc kubenswrapper[4578]: E1003 12:53:53.987315 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:54.487304375 +0000 UTC m=+170.285776559 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.021522 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.026180 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2nzxc" Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.027749 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n6875" Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.048111 4578 patch_prober.go:28] interesting pod/router-default-5444994796-q8kdg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 12:53:54 crc kubenswrapper[4578]: [-]has-synced failed: reason withheld Oct 03 12:53:54 crc kubenswrapper[4578]: [+]process-running ok Oct 03 12:53:54 crc kubenswrapper[4578]: healthz check failed Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.048158 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q8kdg" podUID="59bf594f-62b6-4a2f-a50d-30b5d8357133" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.088070 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:54 crc kubenswrapper[4578]: E1003 12:53:54.088273 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-03 12:53:54.588246529 +0000 UTC m=+170.386718713 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.088350 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:54 crc kubenswrapper[4578]: E1003 12:53:54.088681 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-03 12:53:54.588665002 +0000 UTC m=+170.387137186 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5xxhk" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.102189 4578 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-03T12:53:53.357085214Z","Handler":null,"Name":""} Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.120057 4578 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.120091 4578 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.189614 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.207123 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.251242 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-q8986"] Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.291338 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.299710 4578 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.299752 4578 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.411797 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5xxhk\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.562676 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2nzxc"] Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.685701 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.686386 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.697954 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.698167 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.698384 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.702966 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:54 crc kubenswrapper[4578]: W1003 12:53:54.802372 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4754126b_ff64_44ce_829d_6f8307eed3a7.slice/crio-7f11110b1b1ed7e23e680abe0a193abecf8ee0abbff7963983d8dc16ca42ed30 WatchSource:0}: Error finding container 7f11110b1b1ed7e23e680abe0a193abecf8ee0abbff7963983d8dc16ca42ed30: Status 404 returned error can't find the container with id 7f11110b1b1ed7e23e680abe0a193abecf8ee0abbff7963983d8dc16ca42ed30 Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.814068 4578 generic.go:334] "Generic (PLEG): container finished" podID="fdfd4313-d05e-4cbd-8520-3f929675c662" containerID="05ee2bd4e95415aa471a093fa842b223d3543a457c885f2d0816faaf04cf5fd6" exitCode=0 Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.814141 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4" event={"ID":"fdfd4313-d05e-4cbd-8520-3f929675c662","Type":"ContainerDied","Data":"05ee2bd4e95415aa471a093fa842b223d3543a457c885f2d0816faaf04cf5fd6"} Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.818137 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2242cf82-497d-4e44-bacb-0da986ee2dcb-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2242cf82-497d-4e44-bacb-0da986ee2dcb\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.818210 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2242cf82-497d-4e44-bacb-0da986ee2dcb-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2242cf82-497d-4e44-bacb-0da986ee2dcb\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.821375 4578 generic.go:334] "Generic (PLEG): container finished" podID="fc33b180-1050-403c-81b7-00ffdf579cd7" containerID="b0d50e8d76c0ad56a7adb095f0099c8062d9026c5db226e4ceb765e2d82d19eb" exitCode=0 Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.821432 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdvvk" event={"ID":"fc33b180-1050-403c-81b7-00ffdf579cd7","Type":"ContainerDied","Data":"b0d50e8d76c0ad56a7adb095f0099c8062d9026c5db226e4ceb765e2d82d19eb"} Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.860961 4578 generic.go:334] "Generic (PLEG): container finished" podID="75b24dc1-77ec-4b54-9f85-2443e273eadf" containerID="bbf91797952d16a949c6632514a820d7c4fd06aeb5108932c7905c80d8358b78" exitCode=0 Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.861278 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5hjv" event={"ID":"75b24dc1-77ec-4b54-9f85-2443e273eadf","Type":"ContainerDied","Data":"bbf91797952d16a949c6632514a820d7c4fd06aeb5108932c7905c80d8358b78"} Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.863945 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"30c6876c-90e4-4670-999d-64556e223478","Type":"ContainerStarted","Data":"1592a8610e2949c313cea32fecd73d1b9dbfb7a33e3d7f25742f7eb332b230e8"} Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.870406 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-q8986" event={"ID":"2971c7bb-e0f4-44a8-8f52-9ea8ba199772","Type":"ContainerStarted","Data":"068ce8f04aa971f3b05f279d19df44a5f9be5c3d42f6a379764f5522d35bc81a"} Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.919215 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2242cf82-497d-4e44-bacb-0da986ee2dcb-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2242cf82-497d-4e44-bacb-0da986ee2dcb\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.919255 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2242cf82-497d-4e44-bacb-0da986ee2dcb-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2242cf82-497d-4e44-bacb-0da986ee2dcb\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.919762 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2242cf82-497d-4e44-bacb-0da986ee2dcb-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2242cf82-497d-4e44-bacb-0da986ee2dcb\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.959386 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2242cf82-497d-4e44-bacb-0da986ee2dcb-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2242cf82-497d-4e44-bacb-0da986ee2dcb\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 12:53:54 crc kubenswrapper[4578]: I1003 12:53:54.993850 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 03 12:53:55 crc kubenswrapper[4578]: I1003 12:53:55.013925 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n6875"] Oct 03 12:53:55 crc kubenswrapper[4578]: I1003 12:53:55.034560 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 12:53:55 crc kubenswrapper[4578]: I1003 12:53:55.048542 4578 patch_prober.go:28] interesting pod/router-default-5444994796-q8kdg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 12:53:55 crc kubenswrapper[4578]: [-]has-synced failed: reason withheld Oct 03 12:53:55 crc kubenswrapper[4578]: [+]process-running ok Oct 03 12:53:55 crc kubenswrapper[4578]: healthz check failed Oct 03 12:53:55 crc kubenswrapper[4578]: I1003 12:53:55.048595 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q8kdg" podUID="59bf594f-62b6-4a2f-a50d-30b5d8357133" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 12:53:55 crc kubenswrapper[4578]: W1003 12:53:55.054789 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbeb5b22e_cc9e_453b_bac2_88f7a5d30c29.slice/crio-0cb90b217bc38bb23675b8a5d1974f78860e2786216fe338137591b042899772 WatchSource:0}: Error finding container 0cb90b217bc38bb23675b8a5d1974f78860e2786216fe338137591b042899772: Status 404 returned error can't find the container with id 0cb90b217bc38bb23675b8a5d1974f78860e2786216fe338137591b042899772 Oct 03 12:53:55 crc kubenswrapper[4578]: I1003 12:53:55.558838 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5xxhk"] Oct 03 12:53:55 crc kubenswrapper[4578]: W1003 12:53:55.573528 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a1b400e_f3c1_456c_9d7b_0639a86edbd5.slice/crio-23665e827ecb8df70b8f79e865d560acbb2df4c104887ea3cbf3c3b338579f3d WatchSource:0}: Error finding container 23665e827ecb8df70b8f79e865d560acbb2df4c104887ea3cbf3c3b338579f3d: Status 404 returned error can't find the container with id 23665e827ecb8df70b8f79e865d560acbb2df4c104887ea3cbf3c3b338579f3d Oct 03 12:53:55 crc kubenswrapper[4578]: I1003 12:53:55.596492 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 03 12:53:55 crc kubenswrapper[4578]: I1003 12:53:55.883191 4578 generic.go:334] "Generic (PLEG): container finished" podID="4754126b-ff64-44ce-829d-6f8307eed3a7" containerID="1405176ce613878a2ebb9d581868d9f11a03afb4c45756560dc4a6b475525736" exitCode=0 Oct 03 12:53:55 crc kubenswrapper[4578]: I1003 12:53:55.883248 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nzxc" event={"ID":"4754126b-ff64-44ce-829d-6f8307eed3a7","Type":"ContainerDied","Data":"1405176ce613878a2ebb9d581868d9f11a03afb4c45756560dc4a6b475525736"} Oct 03 12:53:55 crc kubenswrapper[4578]: I1003 12:53:55.883272 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nzxc" event={"ID":"4754126b-ff64-44ce-829d-6f8307eed3a7","Type":"ContainerStarted","Data":"7f11110b1b1ed7e23e680abe0a193abecf8ee0abbff7963983d8dc16ca42ed30"} Oct 03 12:53:55 crc kubenswrapper[4578]: I1003 12:53:55.886176 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"30c6876c-90e4-4670-999d-64556e223478","Type":"ContainerStarted","Data":"7043a3199aa4c8d05303c6562240c9465b990d9c9c839e29ddb737a23a618522"} Oct 03 12:53:55 crc kubenswrapper[4578]: I1003 12:53:55.892588 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" event={"ID":"1a1b400e-f3c1-456c-9d7b-0639a86edbd5","Type":"ContainerStarted","Data":"23665e827ecb8df70b8f79e865d560acbb2df4c104887ea3cbf3c3b338579f3d"} Oct 03 12:53:55 crc kubenswrapper[4578]: I1003 12:53:55.894364 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6875" event={"ID":"beb5b22e-cc9e-453b-bac2-88f7a5d30c29","Type":"ContainerStarted","Data":"0cb90b217bc38bb23675b8a5d1974f78860e2786216fe338137591b042899772"} Oct 03 12:53:55 crc kubenswrapper[4578]: I1003 12:53:55.901975 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2242cf82-497d-4e44-bacb-0da986ee2dcb","Type":"ContainerStarted","Data":"4a00316d02a28083579a2211675e26ac982c960b1a857692e64c0c90a5178921"} Oct 03 12:53:55 crc kubenswrapper[4578]: I1003 12:53:55.902588 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.902486109 podStartE2EDuration="3.902486109s" podCreationTimestamp="2025-10-03 12:53:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:55.898414628 +0000 UTC m=+171.696886812" watchObservedRunningTime="2025-10-03 12:53:55.902486109 +0000 UTC m=+171.700958293" Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.048239 4578 patch_prober.go:28] interesting pod/router-default-5444994796-q8kdg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 12:53:56 crc kubenswrapper[4578]: [-]has-synced failed: reason withheld Oct 03 12:53:56 crc kubenswrapper[4578]: [+]process-running ok Oct 03 12:53:56 crc kubenswrapper[4578]: healthz check failed Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.048496 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q8kdg" podUID="59bf594f-62b6-4a2f-a50d-30b5d8357133" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 12:53:56 crc kubenswrapper[4578]: E1003 12:53:56.058145 4578 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-pod30c6876c_90e4_4670_999d_64556e223478.slice/crio-conmon-7043a3199aa4c8d05303c6562240c9465b990d9c9c839e29ddb737a23a618522.scope\": RecentStats: unable to find data in memory cache]" Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.392184 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4" Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.444159 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fdfd4313-d05e-4cbd-8520-3f929675c662-config-volume\") pod \"fdfd4313-d05e-4cbd-8520-3f929675c662\" (UID: \"fdfd4313-d05e-4cbd-8520-3f929675c662\") " Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.444250 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4cdt\" (UniqueName: \"kubernetes.io/projected/fdfd4313-d05e-4cbd-8520-3f929675c662-kube-api-access-x4cdt\") pod \"fdfd4313-d05e-4cbd-8520-3f929675c662\" (UID: \"fdfd4313-d05e-4cbd-8520-3f929675c662\") " Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.444303 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fdfd4313-d05e-4cbd-8520-3f929675c662-secret-volume\") pod \"fdfd4313-d05e-4cbd-8520-3f929675c662\" (UID: \"fdfd4313-d05e-4cbd-8520-3f929675c662\") " Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.445507 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdfd4313-d05e-4cbd-8520-3f929675c662-config-volume" (OuterVolumeSpecName: "config-volume") pod "fdfd4313-d05e-4cbd-8520-3f929675c662" (UID: "fdfd4313-d05e-4cbd-8520-3f929675c662"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.451125 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdfd4313-d05e-4cbd-8520-3f929675c662-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fdfd4313-d05e-4cbd-8520-3f929675c662" (UID: "fdfd4313-d05e-4cbd-8520-3f929675c662"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.452993 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdfd4313-d05e-4cbd-8520-3f929675c662-kube-api-access-x4cdt" (OuterVolumeSpecName: "kube-api-access-x4cdt") pod "fdfd4313-d05e-4cbd-8520-3f929675c662" (UID: "fdfd4313-d05e-4cbd-8520-3f929675c662"). InnerVolumeSpecName "kube-api-access-x4cdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.545237 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4cdt\" (UniqueName: \"kubernetes.io/projected/fdfd4313-d05e-4cbd-8520-3f929675c662-kube-api-access-x4cdt\") on node \"crc\" DevicePath \"\"" Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.545270 4578 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fdfd4313-d05e-4cbd-8520-3f929675c662-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.545283 4578 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fdfd4313-d05e-4cbd-8520-3f929675c662-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.917097 4578 generic.go:334] "Generic (PLEG): container finished" podID="30c6876c-90e4-4670-999d-64556e223478" containerID="7043a3199aa4c8d05303c6562240c9465b990d9c9c839e29ddb737a23a618522" exitCode=0 Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.926774 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"30c6876c-90e4-4670-999d-64556e223478","Type":"ContainerDied","Data":"7043a3199aa4c8d05303c6562240c9465b990d9c9c839e29ddb737a23a618522"} Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.935471 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-q8986" event={"ID":"2971c7bb-e0f4-44a8-8f52-9ea8ba199772","Type":"ContainerStarted","Data":"e47cae09208e0683f591212e6de6ab95db153b5a44047a822a8e802000a4665d"} Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.967486 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" event={"ID":"1a1b400e-f3c1-456c-9d7b-0639a86edbd5","Type":"ContainerStarted","Data":"2c4a1cd87572a4535793447207bc93408844637dfbc08983920f83393b06d301"} Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.967649 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.978413 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4" event={"ID":"fdfd4313-d05e-4cbd-8520-3f929675c662","Type":"ContainerDied","Data":"d48734357a36fe70d9e2e9cd7a2761d3b9aab129eb319e996fb8e83858c1d90d"} Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.978452 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d48734357a36fe70d9e2e9cd7a2761d3b9aab129eb319e996fb8e83858c1d90d" Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.978545 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4" Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.981818 4578 generic.go:334] "Generic (PLEG): container finished" podID="beb5b22e-cc9e-453b-bac2-88f7a5d30c29" containerID="8f2dcd8ac4499b35a9c3fbdb82d3d415bb581b13a49e033da40e7f26f1567cc0" exitCode=0 Oct 03 12:53:56 crc kubenswrapper[4578]: I1003 12:53:56.981892 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6875" event={"ID":"beb5b22e-cc9e-453b-bac2-88f7a5d30c29","Type":"ContainerDied","Data":"8f2dcd8ac4499b35a9c3fbdb82d3d415bb581b13a49e033da40e7f26f1567cc0"} Oct 03 12:53:57 crc kubenswrapper[4578]: I1003 12:53:57.022077 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" podStartSLOduration=146.022054031 podStartE2EDuration="2m26.022054031s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:56.991516747 +0000 UTC m=+172.789988931" watchObservedRunningTime="2025-10-03 12:53:57.022054031 +0000 UTC m=+172.820526215" Oct 03 12:53:57 crc kubenswrapper[4578]: I1003 12:53:57.022285 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2242cf82-497d-4e44-bacb-0da986ee2dcb","Type":"ContainerStarted","Data":"280fba1dbf093885b47f05af0a047accfc3152d4a20b0135dfdbec8268e928c6"} Oct 03 12:53:57 crc kubenswrapper[4578]: I1003 12:53:57.049413 4578 patch_prober.go:28] interesting pod/router-default-5444994796-q8kdg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 12:53:57 crc kubenswrapper[4578]: [-]has-synced failed: reason withheld Oct 03 12:53:57 crc kubenswrapper[4578]: [+]process-running ok Oct 03 12:53:57 crc kubenswrapper[4578]: healthz check failed Oct 03 12:53:57 crc kubenswrapper[4578]: I1003 12:53:57.049497 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q8kdg" podUID="59bf594f-62b6-4a2f-a50d-30b5d8357133" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 12:53:57 crc kubenswrapper[4578]: I1003 12:53:57.061489 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.061468602 podStartE2EDuration="3.061468602s" podCreationTimestamp="2025-10-03 12:53:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:57.058311119 +0000 UTC m=+172.856783323" watchObservedRunningTime="2025-10-03 12:53:57.061468602 +0000 UTC m=+172.859940786" Oct 03 12:53:57 crc kubenswrapper[4578]: I1003 12:53:57.399834 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:57 crc kubenswrapper[4578]: I1003 12:53:57.408065 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-wmkdd" Oct 03 12:53:58 crc kubenswrapper[4578]: I1003 12:53:58.049152 4578 patch_prober.go:28] interesting pod/router-default-5444994796-q8kdg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 03 12:53:58 crc kubenswrapper[4578]: [-]has-synced failed: reason withheld Oct 03 12:53:58 crc kubenswrapper[4578]: [+]process-running ok Oct 03 12:53:58 crc kubenswrapper[4578]: healthz check failed Oct 03 12:53:58 crc kubenswrapper[4578]: I1003 12:53:58.049218 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q8kdg" podUID="59bf594f-62b6-4a2f-a50d-30b5d8357133" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 03 12:53:58 crc kubenswrapper[4578]: I1003 12:53:58.077942 4578 generic.go:334] "Generic (PLEG): container finished" podID="2242cf82-497d-4e44-bacb-0da986ee2dcb" containerID="280fba1dbf093885b47f05af0a047accfc3152d4a20b0135dfdbec8268e928c6" exitCode=0 Oct 03 12:53:58 crc kubenswrapper[4578]: I1003 12:53:58.078021 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2242cf82-497d-4e44-bacb-0da986ee2dcb","Type":"ContainerDied","Data":"280fba1dbf093885b47f05af0a047accfc3152d4a20b0135dfdbec8268e928c6"} Oct 03 12:53:58 crc kubenswrapper[4578]: I1003 12:53:58.092340 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-q8986" event={"ID":"2971c7bb-e0f4-44a8-8f52-9ea8ba199772","Type":"ContainerStarted","Data":"f33f1b9a2c44256bcc02ec4cf74123159633d4005150e122124d01272e7bb242"} Oct 03 12:53:58 crc kubenswrapper[4578]: I1003 12:53:58.150051 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-q8986" podStartSLOduration=147.150036762 podStartE2EDuration="2m27.150036762s" podCreationTimestamp="2025-10-03 12:51:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:53:58.140449553 +0000 UTC m=+173.938921757" watchObservedRunningTime="2025-10-03 12:53:58.150036762 +0000 UTC m=+173.948508946" Oct 03 12:53:58 crc kubenswrapper[4578]: I1003 12:53:58.499106 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-z7c9x" Oct 03 12:53:58 crc kubenswrapper[4578]: I1003 12:53:58.572678 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 12:53:58 crc kubenswrapper[4578]: I1003 12:53:58.701197 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/30c6876c-90e4-4670-999d-64556e223478-kubelet-dir\") pod \"30c6876c-90e4-4670-999d-64556e223478\" (UID: \"30c6876c-90e4-4670-999d-64556e223478\") " Oct 03 12:53:58 crc kubenswrapper[4578]: I1003 12:53:58.701296 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30c6876c-90e4-4670-999d-64556e223478-kube-api-access\") pod \"30c6876c-90e4-4670-999d-64556e223478\" (UID: \"30c6876c-90e4-4670-999d-64556e223478\") " Oct 03 12:53:58 crc kubenswrapper[4578]: I1003 12:53:58.702226 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/30c6876c-90e4-4670-999d-64556e223478-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "30c6876c-90e4-4670-999d-64556e223478" (UID: "30c6876c-90e4-4670-999d-64556e223478"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 12:53:58 crc kubenswrapper[4578]: I1003 12:53:58.710914 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30c6876c-90e4-4670-999d-64556e223478-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "30c6876c-90e4-4670-999d-64556e223478" (UID: "30c6876c-90e4-4670-999d-64556e223478"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:53:58 crc kubenswrapper[4578]: I1003 12:53:58.802556 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/30c6876c-90e4-4670-999d-64556e223478-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 03 12:53:58 crc kubenswrapper[4578]: I1003 12:53:58.803216 4578 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/30c6876c-90e4-4670-999d-64556e223478-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 03 12:53:59 crc kubenswrapper[4578]: I1003 12:53:59.053468 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:59 crc kubenswrapper[4578]: I1003 12:53:59.055733 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-q8kdg" Oct 03 12:53:59 crc kubenswrapper[4578]: I1003 12:53:59.175103 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 03 12:53:59 crc kubenswrapper[4578]: I1003 12:53:59.175709 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"30c6876c-90e4-4670-999d-64556e223478","Type":"ContainerDied","Data":"1592a8610e2949c313cea32fecd73d1b9dbfb7a33e3d7f25742f7eb332b230e8"} Oct 03 12:53:59 crc kubenswrapper[4578]: I1003 12:53:59.175731 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1592a8610e2949c313cea32fecd73d1b9dbfb7a33e3d7f25742f7eb332b230e8" Oct 03 12:53:59 crc kubenswrapper[4578]: I1003 12:53:59.584217 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 12:53:59 crc kubenswrapper[4578]: I1003 12:53:59.621298 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2242cf82-497d-4e44-bacb-0da986ee2dcb-kubelet-dir\") pod \"2242cf82-497d-4e44-bacb-0da986ee2dcb\" (UID: \"2242cf82-497d-4e44-bacb-0da986ee2dcb\") " Oct 03 12:53:59 crc kubenswrapper[4578]: I1003 12:53:59.621397 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2242cf82-497d-4e44-bacb-0da986ee2dcb-kube-api-access\") pod \"2242cf82-497d-4e44-bacb-0da986ee2dcb\" (UID: \"2242cf82-497d-4e44-bacb-0da986ee2dcb\") " Oct 03 12:53:59 crc kubenswrapper[4578]: I1003 12:53:59.624075 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2242cf82-497d-4e44-bacb-0da986ee2dcb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2242cf82-497d-4e44-bacb-0da986ee2dcb" (UID: "2242cf82-497d-4e44-bacb-0da986ee2dcb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 12:53:59 crc kubenswrapper[4578]: I1003 12:53:59.634211 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2242cf82-497d-4e44-bacb-0da986ee2dcb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2242cf82-497d-4e44-bacb-0da986ee2dcb" (UID: "2242cf82-497d-4e44-bacb-0da986ee2dcb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:53:59 crc kubenswrapper[4578]: I1003 12:53:59.723043 4578 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2242cf82-497d-4e44-bacb-0da986ee2dcb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 03 12:53:59 crc kubenswrapper[4578]: I1003 12:53:59.723117 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2242cf82-497d-4e44-bacb-0da986ee2dcb-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 03 12:54:00 crc kubenswrapper[4578]: I1003 12:54:00.190903 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2242cf82-497d-4e44-bacb-0da986ee2dcb","Type":"ContainerDied","Data":"4a00316d02a28083579a2211675e26ac982c960b1a857692e64c0c90a5178921"} Oct 03 12:54:00 crc kubenswrapper[4578]: I1003 12:54:00.190939 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a00316d02a28083579a2211675e26ac982c960b1a857692e64c0c90a5178921" Oct 03 12:54:00 crc kubenswrapper[4578]: I1003 12:54:00.190972 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 03 12:54:02 crc kubenswrapper[4578]: I1003 12:54:02.251978 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:54:02 crc kubenswrapper[4578]: I1003 12:54:02.255145 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-ns26l" Oct 03 12:54:02 crc kubenswrapper[4578]: I1003 12:54:02.284508 4578 patch_prober.go:28] interesting pod/downloads-7954f5f757-d2rzv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 03 12:54:02 crc kubenswrapper[4578]: I1003 12:54:02.284560 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-d2rzv" podUID="de78c746-e145-48ba-8601-a3ff2c6b2b62" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 03 12:54:02 crc kubenswrapper[4578]: I1003 12:54:02.285004 4578 patch_prober.go:28] interesting pod/downloads-7954f5f757-d2rzv container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 03 12:54:02 crc kubenswrapper[4578]: I1003 12:54:02.285048 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-d2rzv" podUID="de78c746-e145-48ba-8601-a3ff2c6b2b62" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.29:8080/\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 03 12:54:05 crc kubenswrapper[4578]: I1003 12:54:05.091807 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 12:54:05 crc kubenswrapper[4578]: I1003 12:54:05.092136 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 12:54:10 crc kubenswrapper[4578]: I1003 12:54:10.103936 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 03 12:54:12 crc kubenswrapper[4578]: I1003 12:54:12.291160 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-d2rzv" Oct 03 12:54:14 crc kubenswrapper[4578]: I1003 12:54:14.708077 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 12:54:23 crc kubenswrapper[4578]: I1003 12:54:23.477985 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-rb78z" Oct 03 12:54:26 crc kubenswrapper[4578]: E1003 12:54:26.135093 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 03 12:54:26 crc kubenswrapper[4578]: E1003 12:54:26.135566 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pr5cg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-2nzxc_openshift-marketplace(4754126b-ff64-44ce-829d-6f8307eed3a7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 12:54:26 crc kubenswrapper[4578]: E1003 12:54:26.137073 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-2nzxc" podUID="4754126b-ff64-44ce-829d-6f8307eed3a7" Oct 03 12:54:29 crc kubenswrapper[4578]: E1003 12:54:29.015043 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 03 12:54:29 crc kubenswrapper[4578]: E1003 12:54:29.015588 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rfh44,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-n6875_openshift-marketplace(beb5b22e-cc9e-453b-bac2-88f7a5d30c29): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 12:54:29 crc kubenswrapper[4578]: E1003 12:54:29.016868 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-n6875" podUID="beb5b22e-cc9e-453b-bac2-88f7a5d30c29" Oct 03 12:54:30 crc kubenswrapper[4578]: E1003 12:54:30.384233 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-n6875" podUID="beb5b22e-cc9e-453b-bac2-88f7a5d30c29" Oct 03 12:54:30 crc kubenswrapper[4578]: E1003 12:54:30.446887 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 03 12:54:30 crc kubenswrapper[4578]: E1003 12:54:30.447480 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-55rzg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-k4z8f_openshift-marketplace(445697f9-cf9b-44d0-b76a-c75e6910b758): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 12:54:30 crc kubenswrapper[4578]: E1003 12:54:30.449192 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-k4z8f" podUID="445697f9-cf9b-44d0-b76a-c75e6910b758" Oct 03 12:54:31 crc kubenswrapper[4578]: E1003 12:54:31.976417 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-k4z8f" podUID="445697f9-cf9b-44d0-b76a-c75e6910b758" Oct 03 12:54:32 crc kubenswrapper[4578]: E1003 12:54:32.064919 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 03 12:54:32 crc kubenswrapper[4578]: E1003 12:54:32.065088 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ws6gc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-94bd2_openshift-marketplace(2244329a-bbb1-4c5a-a170-55106dcb21c2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 12:54:32 crc kubenswrapper[4578]: E1003 12:54:32.066374 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-94bd2" podUID="2244329a-bbb1-4c5a-a170-55106dcb21c2" Oct 03 12:54:32 crc kubenswrapper[4578]: E1003 12:54:32.071787 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 03 12:54:32 crc kubenswrapper[4578]: E1003 12:54:32.072011 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t287h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-ldfq4_openshift-marketplace(b40b7869-c4f0-427d-b99c-fca7f0d8b937): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 12:54:32 crc kubenswrapper[4578]: E1003 12:54:32.073247 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-ldfq4" podUID="b40b7869-c4f0-427d-b99c-fca7f0d8b937" Oct 03 12:54:34 crc kubenswrapper[4578]: E1003 12:54:34.950668 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-94bd2" podUID="2244329a-bbb1-4c5a-a170-55106dcb21c2" Oct 03 12:54:34 crc kubenswrapper[4578]: E1003 12:54:34.951834 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-ldfq4" podUID="b40b7869-c4f0-427d-b99c-fca7f0d8b937" Oct 03 12:54:35 crc kubenswrapper[4578]: E1003 12:54:35.028125 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 03 12:54:35 crc kubenswrapper[4578]: E1003 12:54:35.028299 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7rccr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-hmk8t_openshift-marketplace(742a4571-0a86-4cb4-bb93-8b46219f6d49): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 12:54:35 crc kubenswrapper[4578]: E1003 12:54:35.029595 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-hmk8t" podUID="742a4571-0a86-4cb4-bb93-8b46219f6d49" Oct 03 12:54:35 crc kubenswrapper[4578]: I1003 12:54:35.092025 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 12:54:35 crc kubenswrapper[4578]: I1003 12:54:35.092098 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 12:54:35 crc kubenswrapper[4578]: I1003 12:54:35.092144 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 12:54:35 crc kubenswrapper[4578]: I1003 12:54:35.092936 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716"} pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 12:54:35 crc kubenswrapper[4578]: I1003 12:54:35.093087 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" containerID="cri-o://54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716" gracePeriod=600 Oct 03 12:54:35 crc kubenswrapper[4578]: I1003 12:54:35.406884 4578 generic.go:334] "Generic (PLEG): container finished" podID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerID="54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716" exitCode=0 Oct 03 12:54:35 crc kubenswrapper[4578]: I1003 12:54:35.407216 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerDied","Data":"54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716"} Oct 03 12:54:35 crc kubenswrapper[4578]: E1003 12:54:35.624125 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-hmk8t" podUID="742a4571-0a86-4cb4-bb93-8b46219f6d49" Oct 03 12:54:35 crc kubenswrapper[4578]: E1003 12:54:35.684125 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 03 12:54:35 crc kubenswrapper[4578]: E1003 12:54:35.684428 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s5bqk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-w5hjv_openshift-marketplace(75b24dc1-77ec-4b54-9f85-2443e273eadf): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 12:54:35 crc kubenswrapper[4578]: E1003 12:54:35.685582 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-w5hjv" podUID="75b24dc1-77ec-4b54-9f85-2443e273eadf" Oct 03 12:54:35 crc kubenswrapper[4578]: E1003 12:54:35.712529 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 03 12:54:35 crc kubenswrapper[4578]: E1003 12:54:35.712675 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xmvw5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-qdvvk_openshift-marketplace(fc33b180-1050-403c-81b7-00ffdf579cd7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 12:54:35 crc kubenswrapper[4578]: E1003 12:54:35.713832 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-qdvvk" podUID="fc33b180-1050-403c-81b7-00ffdf579cd7" Oct 03 12:54:36 crc kubenswrapper[4578]: I1003 12:54:36.412889 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"199792b4b113133fd7bee4bb1715ec98bb415c3fcc2500bb9a60f6c23d31f8a5"} Oct 03 12:54:36 crc kubenswrapper[4578]: E1003 12:54:36.414882 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-qdvvk" podUID="fc33b180-1050-403c-81b7-00ffdf579cd7" Oct 03 12:54:36 crc kubenswrapper[4578]: E1003 12:54:36.415142 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-w5hjv" podUID="75b24dc1-77ec-4b54-9f85-2443e273eadf" Oct 03 12:54:38 crc kubenswrapper[4578]: I1003 12:54:38.423548 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nzxc" event={"ID":"4754126b-ff64-44ce-829d-6f8307eed3a7","Type":"ContainerStarted","Data":"8e5cc0d204697d10222adbe7b13e77da76e9901228d254392428527423906c49"} Oct 03 12:54:39 crc kubenswrapper[4578]: I1003 12:54:39.428808 4578 generic.go:334] "Generic (PLEG): container finished" podID="4754126b-ff64-44ce-829d-6f8307eed3a7" containerID="8e5cc0d204697d10222adbe7b13e77da76e9901228d254392428527423906c49" exitCode=0 Oct 03 12:54:39 crc kubenswrapper[4578]: I1003 12:54:39.428858 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nzxc" event={"ID":"4754126b-ff64-44ce-829d-6f8307eed3a7","Type":"ContainerDied","Data":"8e5cc0d204697d10222adbe7b13e77da76e9901228d254392428527423906c49"} Oct 03 12:54:40 crc kubenswrapper[4578]: I1003 12:54:40.435598 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nzxc" event={"ID":"4754126b-ff64-44ce-829d-6f8307eed3a7","Type":"ContainerStarted","Data":"ece99ca601694b6dc522a9b31141802358143d67b4dfdbd69255812042b7a6d7"} Oct 03 12:54:40 crc kubenswrapper[4578]: I1003 12:54:40.468326 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2nzxc" podStartSLOduration=4.640750451 podStartE2EDuration="47.468300689s" podCreationTimestamp="2025-10-03 12:53:53 +0000 UTC" firstStartedPulling="2025-10-03 12:53:57.033164379 +0000 UTC m=+172.831636563" lastFinishedPulling="2025-10-03 12:54:39.860714607 +0000 UTC m=+215.659186801" observedRunningTime="2025-10-03 12:54:40.465606572 +0000 UTC m=+216.264078756" watchObservedRunningTime="2025-10-03 12:54:40.468300689 +0000 UTC m=+216.266772883" Oct 03 12:54:44 crc kubenswrapper[4578]: I1003 12:54:44.027033 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2nzxc" Oct 03 12:54:44 crc kubenswrapper[4578]: I1003 12:54:44.027404 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2nzxc" Oct 03 12:54:45 crc kubenswrapper[4578]: I1003 12:54:45.183032 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2nzxc" podUID="4754126b-ff64-44ce-829d-6f8307eed3a7" containerName="registry-server" probeResult="failure" output=< Oct 03 12:54:45 crc kubenswrapper[4578]: timeout: failed to connect service ":50051" within 1s Oct 03 12:54:45 crc kubenswrapper[4578]: > Oct 03 12:54:45 crc kubenswrapper[4578]: I1003 12:54:45.457830 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6875" event={"ID":"beb5b22e-cc9e-453b-bac2-88f7a5d30c29","Type":"ContainerStarted","Data":"c5a5f2cbc2c1b9dbbf9f69af41dc4e19a2d2abae656b50a51f469220375645f2"} Oct 03 12:54:46 crc kubenswrapper[4578]: I1003 12:54:46.465199 4578 generic.go:334] "Generic (PLEG): container finished" podID="beb5b22e-cc9e-453b-bac2-88f7a5d30c29" containerID="c5a5f2cbc2c1b9dbbf9f69af41dc4e19a2d2abae656b50a51f469220375645f2" exitCode=0 Oct 03 12:54:46 crc kubenswrapper[4578]: I1003 12:54:46.465245 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6875" event={"ID":"beb5b22e-cc9e-453b-bac2-88f7a5d30c29","Type":"ContainerDied","Data":"c5a5f2cbc2c1b9dbbf9f69af41dc4e19a2d2abae656b50a51f469220375645f2"} Oct 03 12:54:54 crc kubenswrapper[4578]: I1003 12:54:54.248964 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2nzxc" Oct 03 12:54:54 crc kubenswrapper[4578]: I1003 12:54:54.304294 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2nzxc" Oct 03 12:54:55 crc kubenswrapper[4578]: I1003 12:54:55.145810 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2nzxc"] Oct 03 12:54:55 crc kubenswrapper[4578]: I1003 12:54:55.512720 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2nzxc" podUID="4754126b-ff64-44ce-829d-6f8307eed3a7" containerName="registry-server" containerID="cri-o://ece99ca601694b6dc522a9b31141802358143d67b4dfdbd69255812042b7a6d7" gracePeriod=2 Oct 03 12:54:56 crc kubenswrapper[4578]: I1003 12:54:56.518817 4578 generic.go:334] "Generic (PLEG): container finished" podID="4754126b-ff64-44ce-829d-6f8307eed3a7" containerID="ece99ca601694b6dc522a9b31141802358143d67b4dfdbd69255812042b7a6d7" exitCode=0 Oct 03 12:54:56 crc kubenswrapper[4578]: I1003 12:54:56.518888 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nzxc" event={"ID":"4754126b-ff64-44ce-829d-6f8307eed3a7","Type":"ContainerDied","Data":"ece99ca601694b6dc522a9b31141802358143d67b4dfdbd69255812042b7a6d7"} Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.012050 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2nzxc" Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.116450 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pr5cg\" (UniqueName: \"kubernetes.io/projected/4754126b-ff64-44ce-829d-6f8307eed3a7-kube-api-access-pr5cg\") pod \"4754126b-ff64-44ce-829d-6f8307eed3a7\" (UID: \"4754126b-ff64-44ce-829d-6f8307eed3a7\") " Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.116550 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4754126b-ff64-44ce-829d-6f8307eed3a7-catalog-content\") pod \"4754126b-ff64-44ce-829d-6f8307eed3a7\" (UID: \"4754126b-ff64-44ce-829d-6f8307eed3a7\") " Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.116582 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4754126b-ff64-44ce-829d-6f8307eed3a7-utilities\") pod \"4754126b-ff64-44ce-829d-6f8307eed3a7\" (UID: \"4754126b-ff64-44ce-829d-6f8307eed3a7\") " Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.123859 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4754126b-ff64-44ce-829d-6f8307eed3a7-utilities" (OuterVolumeSpecName: "utilities") pod "4754126b-ff64-44ce-829d-6f8307eed3a7" (UID: "4754126b-ff64-44ce-829d-6f8307eed3a7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.142817 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4754126b-ff64-44ce-829d-6f8307eed3a7-kube-api-access-pr5cg" (OuterVolumeSpecName: "kube-api-access-pr5cg") pod "4754126b-ff64-44ce-829d-6f8307eed3a7" (UID: "4754126b-ff64-44ce-829d-6f8307eed3a7"). InnerVolumeSpecName "kube-api-access-pr5cg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.218444 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pr5cg\" (UniqueName: \"kubernetes.io/projected/4754126b-ff64-44ce-829d-6f8307eed3a7-kube-api-access-pr5cg\") on node \"crc\" DevicePath \"\"" Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.218477 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4754126b-ff64-44ce-829d-6f8307eed3a7-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.246233 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4754126b-ff64-44ce-829d-6f8307eed3a7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4754126b-ff64-44ce-829d-6f8307eed3a7" (UID: "4754126b-ff64-44ce-829d-6f8307eed3a7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.320076 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4754126b-ff64-44ce-829d-6f8307eed3a7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.530308 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nzxc" event={"ID":"4754126b-ff64-44ce-829d-6f8307eed3a7","Type":"ContainerDied","Data":"7f11110b1b1ed7e23e680abe0a193abecf8ee0abbff7963983d8dc16ca42ed30"} Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.530355 4578 scope.go:117] "RemoveContainer" containerID="ece99ca601694b6dc522a9b31141802358143d67b4dfdbd69255812042b7a6d7" Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.530455 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2nzxc" Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.537120 4578 generic.go:334] "Generic (PLEG): container finished" podID="742a4571-0a86-4cb4-bb93-8b46219f6d49" containerID="35e099489bdb34a5f63b43d43b657ee4608e45d458c23deab329cc34f1f8fde0" exitCode=0 Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.537196 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hmk8t" event={"ID":"742a4571-0a86-4cb4-bb93-8b46219f6d49","Type":"ContainerDied","Data":"35e099489bdb34a5f63b43d43b657ee4608e45d458c23deab329cc34f1f8fde0"} Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.547996 4578 generic.go:334] "Generic (PLEG): container finished" podID="2244329a-bbb1-4c5a-a170-55106dcb21c2" containerID="7ddf64ec0e9a781add1eaebdcb55208d8b10f4aa18b5ff52613b5dc099126f54" exitCode=0 Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.548058 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-94bd2" event={"ID":"2244329a-bbb1-4c5a-a170-55106dcb21c2","Type":"ContainerDied","Data":"7ddf64ec0e9a781add1eaebdcb55208d8b10f4aa18b5ff52613b5dc099126f54"} Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.553282 4578 generic.go:334] "Generic (PLEG): container finished" podID="b40b7869-c4f0-427d-b99c-fca7f0d8b937" containerID="2e071a1b027ff081eb45821e9388f826ec1b24a2089965350e316344a307e6d9" exitCode=0 Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.553362 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ldfq4" event={"ID":"b40b7869-c4f0-427d-b99c-fca7f0d8b937","Type":"ContainerDied","Data":"2e071a1b027ff081eb45821e9388f826ec1b24a2089965350e316344a307e6d9"} Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.560778 4578 generic.go:334] "Generic (PLEG): container finished" podID="fc33b180-1050-403c-81b7-00ffdf579cd7" containerID="045dd477c0852a3a82603181c21fe9f46db6f6d09f5f0f54d58c0770bae6cddd" exitCode=0 Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.560875 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdvvk" event={"ID":"fc33b180-1050-403c-81b7-00ffdf579cd7","Type":"ContainerDied","Data":"045dd477c0852a3a82603181c21fe9f46db6f6d09f5f0f54d58c0770bae6cddd"} Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.565264 4578 generic.go:334] "Generic (PLEG): container finished" podID="75b24dc1-77ec-4b54-9f85-2443e273eadf" containerID="5ef68c2cbc6ffb046975d4b63e43959fee1d0d2899a5b0a8878d2c1f40c51f76" exitCode=0 Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.565370 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5hjv" event={"ID":"75b24dc1-77ec-4b54-9f85-2443e273eadf","Type":"ContainerDied","Data":"5ef68c2cbc6ffb046975d4b63e43959fee1d0d2899a5b0a8878d2c1f40c51f76"} Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.577707 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6875" event={"ID":"beb5b22e-cc9e-453b-bac2-88f7a5d30c29","Type":"ContainerStarted","Data":"d6dfbd083ed07317b9c09ccde3afa3a6e322e7bab844def4d21e504843685d41"} Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.579661 4578 generic.go:334] "Generic (PLEG): container finished" podID="445697f9-cf9b-44d0-b76a-c75e6910b758" containerID="99d2098662c456703a44270cb86d8f93b1d295f27ffa31a25d5954469f55001d" exitCode=0 Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.579683 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4z8f" event={"ID":"445697f9-cf9b-44d0-b76a-c75e6910b758","Type":"ContainerDied","Data":"99d2098662c456703a44270cb86d8f93b1d295f27ffa31a25d5954469f55001d"} Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.607167 4578 scope.go:117] "RemoveContainer" containerID="8e5cc0d204697d10222adbe7b13e77da76e9901228d254392428527423906c49" Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.636955 4578 scope.go:117] "RemoveContainer" containerID="1405176ce613878a2ebb9d581868d9f11a03afb4c45756560dc4a6b475525736" Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.642014 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2nzxc"] Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.647383 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2nzxc"] Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.692415 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n6875" podStartSLOduration=5.926479771 podStartE2EDuration="1m6.692396714s" podCreationTimestamp="2025-10-03 12:53:52 +0000 UTC" firstStartedPulling="2025-10-03 12:53:56.990797004 +0000 UTC m=+172.789269188" lastFinishedPulling="2025-10-03 12:54:57.756713927 +0000 UTC m=+233.555186131" observedRunningTime="2025-10-03 12:54:58.690793402 +0000 UTC m=+234.489265586" watchObservedRunningTime="2025-10-03 12:54:58.692396714 +0000 UTC m=+234.490868898" Oct 03 12:54:58 crc kubenswrapper[4578]: I1003 12:54:58.917173 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4754126b-ff64-44ce-829d-6f8307eed3a7" path="/var/lib/kubelet/pods/4754126b-ff64-44ce-829d-6f8307eed3a7/volumes" Oct 03 12:54:59 crc kubenswrapper[4578]: I1003 12:54:59.586048 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5hjv" event={"ID":"75b24dc1-77ec-4b54-9f85-2443e273eadf","Type":"ContainerStarted","Data":"c0443a230b0f1a59a152fd9fe5e5afab476c4c195bdbe35c0a33446f5e491377"} Oct 03 12:54:59 crc kubenswrapper[4578]: I1003 12:54:59.588244 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4z8f" event={"ID":"445697f9-cf9b-44d0-b76a-c75e6910b758","Type":"ContainerStarted","Data":"defd005cc78445af279f2734b9183efb0d9288da3432f00d4abad565e0ac6c99"} Oct 03 12:54:59 crc kubenswrapper[4578]: I1003 12:54:59.591060 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hmk8t" event={"ID":"742a4571-0a86-4cb4-bb93-8b46219f6d49","Type":"ContainerStarted","Data":"42b16065e5cd5959e35719370e31b905f5279cd890408d7c5543fa49e060d5a5"} Oct 03 12:54:59 crc kubenswrapper[4578]: I1003 12:54:59.593145 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-94bd2" event={"ID":"2244329a-bbb1-4c5a-a170-55106dcb21c2","Type":"ContainerStarted","Data":"0f7c32f6c75cc7ee3ee180c5115c82b1062f762be4bf0720ab7c446af7f634b3"} Oct 03 12:54:59 crc kubenswrapper[4578]: I1003 12:54:59.595273 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ldfq4" event={"ID":"b40b7869-c4f0-427d-b99c-fca7f0d8b937","Type":"ContainerStarted","Data":"0eb7978f73b5b71a41f58d92e0989daa504e621cc70de06a65d413e9353b45bb"} Oct 03 12:54:59 crc kubenswrapper[4578]: I1003 12:54:59.598252 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdvvk" event={"ID":"fc33b180-1050-403c-81b7-00ffdf579cd7","Type":"ContainerStarted","Data":"f5f0401a8724dfcf17f3d9626ddd580c1e9ee118d0262811f78d8c4842c8ad19"} Oct 03 12:54:59 crc kubenswrapper[4578]: I1003 12:54:59.608878 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w5hjv" podStartSLOduration=3.198171852 podStartE2EDuration="1m8.60886273s" podCreationTimestamp="2025-10-03 12:53:51 +0000 UTC" firstStartedPulling="2025-10-03 12:53:53.606142901 +0000 UTC m=+169.404615085" lastFinishedPulling="2025-10-03 12:54:59.016833779 +0000 UTC m=+234.815305963" observedRunningTime="2025-10-03 12:54:59.606719831 +0000 UTC m=+235.405192015" watchObservedRunningTime="2025-10-03 12:54:59.60886273 +0000 UTC m=+235.407334914" Oct 03 12:54:59 crc kubenswrapper[4578]: I1003 12:54:59.632436 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ldfq4" podStartSLOduration=3.925332298 podStartE2EDuration="1m10.632417399s" podCreationTimestamp="2025-10-03 12:53:49 +0000 UTC" firstStartedPulling="2025-10-03 12:53:52.392752384 +0000 UTC m=+168.191224568" lastFinishedPulling="2025-10-03 12:54:59.099837485 +0000 UTC m=+234.898309669" observedRunningTime="2025-10-03 12:54:59.629264308 +0000 UTC m=+235.427736502" watchObservedRunningTime="2025-10-03 12:54:59.632417399 +0000 UTC m=+235.430889593" Oct 03 12:54:59 crc kubenswrapper[4578]: I1003 12:54:59.670665 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qdvvk" podStartSLOduration=3.567160812 podStartE2EDuration="1m7.67062307s" podCreationTimestamp="2025-10-03 12:53:52 +0000 UTC" firstStartedPulling="2025-10-03 12:53:54.878071024 +0000 UTC m=+170.676543208" lastFinishedPulling="2025-10-03 12:54:58.981533282 +0000 UTC m=+234.780005466" observedRunningTime="2025-10-03 12:54:59.65198074 +0000 UTC m=+235.450452924" watchObservedRunningTime="2025-10-03 12:54:59.67062307 +0000 UTC m=+235.469095254" Oct 03 12:54:59 crc kubenswrapper[4578]: I1003 12:54:59.693954 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hmk8t" podStartSLOduration=3.942374468 podStartE2EDuration="1m10.693938682s" podCreationTimestamp="2025-10-03 12:53:49 +0000 UTC" firstStartedPulling="2025-10-03 12:53:52.356109833 +0000 UTC m=+168.154582017" lastFinishedPulling="2025-10-03 12:54:59.107674047 +0000 UTC m=+234.906146231" observedRunningTime="2025-10-03 12:54:59.693575131 +0000 UTC m=+235.492047335" watchObservedRunningTime="2025-10-03 12:54:59.693938682 +0000 UTC m=+235.492410866" Oct 03 12:54:59 crc kubenswrapper[4578]: I1003 12:54:59.695986 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k4z8f" podStartSLOduration=4.07089643 podStartE2EDuration="1m10.695976098s" podCreationTimestamp="2025-10-03 12:53:49 +0000 UTC" firstStartedPulling="2025-10-03 12:53:52.543918116 +0000 UTC m=+168.342390300" lastFinishedPulling="2025-10-03 12:54:59.168997784 +0000 UTC m=+234.967469968" observedRunningTime="2025-10-03 12:54:59.673288557 +0000 UTC m=+235.471760741" watchObservedRunningTime="2025-10-03 12:54:59.695976098 +0000 UTC m=+235.494448282" Oct 03 12:54:59 crc kubenswrapper[4578]: I1003 12:54:59.712864 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-94bd2" podStartSLOduration=3.017525529 podStartE2EDuration="1m9.712843051s" podCreationTimestamp="2025-10-03 12:53:50 +0000 UTC" firstStartedPulling="2025-10-03 12:53:52.382136022 +0000 UTC m=+168.180608206" lastFinishedPulling="2025-10-03 12:54:59.077453544 +0000 UTC m=+234.875925728" observedRunningTime="2025-10-03 12:54:59.711967313 +0000 UTC m=+235.510439497" watchObservedRunningTime="2025-10-03 12:54:59.712843051 +0000 UTC m=+235.511315235" Oct 03 12:54:59 crc kubenswrapper[4578]: I1003 12:54:59.940288 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ldfq4" Oct 03 12:54:59 crc kubenswrapper[4578]: I1003 12:54:59.940611 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ldfq4" Oct 03 12:55:00 crc kubenswrapper[4578]: I1003 12:55:00.104020 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hmk8t" Oct 03 12:55:00 crc kubenswrapper[4578]: I1003 12:55:00.104080 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hmk8t" Oct 03 12:55:00 crc kubenswrapper[4578]: I1003 12:55:00.333845 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k4z8f" Oct 03 12:55:00 crc kubenswrapper[4578]: I1003 12:55:00.333923 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k4z8f" Oct 03 12:55:00 crc kubenswrapper[4578]: I1003 12:55:00.858266 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-94bd2" Oct 03 12:55:00 crc kubenswrapper[4578]: I1003 12:55:00.858582 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-94bd2" Oct 03 12:55:01 crc kubenswrapper[4578]: I1003 12:55:01.004319 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-ldfq4" podUID="b40b7869-c4f0-427d-b99c-fca7f0d8b937" containerName="registry-server" probeResult="failure" output=< Oct 03 12:55:01 crc kubenswrapper[4578]: timeout: failed to connect service ":50051" within 1s Oct 03 12:55:01 crc kubenswrapper[4578]: > Oct 03 12:55:01 crc kubenswrapper[4578]: I1003 12:55:01.175955 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-hmk8t" podUID="742a4571-0a86-4cb4-bb93-8b46219f6d49" containerName="registry-server" probeResult="failure" output=< Oct 03 12:55:01 crc kubenswrapper[4578]: timeout: failed to connect service ":50051" within 1s Oct 03 12:55:01 crc kubenswrapper[4578]: > Oct 03 12:55:01 crc kubenswrapper[4578]: I1003 12:55:01.382409 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-k4z8f" podUID="445697f9-cf9b-44d0-b76a-c75e6910b758" containerName="registry-server" probeResult="failure" output=< Oct 03 12:55:01 crc kubenswrapper[4578]: timeout: failed to connect service ":50051" within 1s Oct 03 12:55:01 crc kubenswrapper[4578]: > Oct 03 12:55:01 crc kubenswrapper[4578]: I1003 12:55:01.896617 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-94bd2" podUID="2244329a-bbb1-4c5a-a170-55106dcb21c2" containerName="registry-server" probeResult="failure" output=< Oct 03 12:55:01 crc kubenswrapper[4578]: timeout: failed to connect service ":50051" within 1s Oct 03 12:55:01 crc kubenswrapper[4578]: > Oct 03 12:55:02 crc kubenswrapper[4578]: I1003 12:55:02.109392 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w5hjv" Oct 03 12:55:02 crc kubenswrapper[4578]: I1003 12:55:02.109438 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w5hjv" Oct 03 12:55:02 crc kubenswrapper[4578]: I1003 12:55:02.154276 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w5hjv" Oct 03 12:55:02 crc kubenswrapper[4578]: I1003 12:55:02.505464 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qdvvk" Oct 03 12:55:02 crc kubenswrapper[4578]: I1003 12:55:02.506045 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qdvvk" Oct 03 12:55:02 crc kubenswrapper[4578]: I1003 12:55:02.547147 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qdvvk" Oct 03 12:55:04 crc kubenswrapper[4578]: I1003 12:55:04.029134 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n6875" Oct 03 12:55:04 crc kubenswrapper[4578]: I1003 12:55:04.029487 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n6875" Oct 03 12:55:04 crc kubenswrapper[4578]: I1003 12:55:04.064495 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n6875" Oct 03 12:55:04 crc kubenswrapper[4578]: I1003 12:55:04.663125 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n6875" Oct 03 12:55:09 crc kubenswrapper[4578]: I1003 12:55:09.983662 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ldfq4" Oct 03 12:55:10 crc kubenswrapper[4578]: I1003 12:55:10.021269 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ldfq4" Oct 03 12:55:10 crc kubenswrapper[4578]: I1003 12:55:10.147166 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hmk8t" Oct 03 12:55:10 crc kubenswrapper[4578]: I1003 12:55:10.192054 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hmk8t" Oct 03 12:55:10 crc kubenswrapper[4578]: I1003 12:55:10.372092 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k4z8f" Oct 03 12:55:10 crc kubenswrapper[4578]: I1003 12:55:10.415046 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k4z8f" Oct 03 12:55:10 crc kubenswrapper[4578]: I1003 12:55:10.918238 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-94bd2" Oct 03 12:55:10 crc kubenswrapper[4578]: I1003 12:55:10.969656 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-94bd2" Oct 03 12:55:12 crc kubenswrapper[4578]: I1003 12:55:12.017704 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-94bd2"] Oct 03 12:55:12 crc kubenswrapper[4578]: I1003 12:55:12.154884 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w5hjv" Oct 03 12:55:12 crc kubenswrapper[4578]: I1003 12:55:12.249997 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7br6x"] Oct 03 12:55:12 crc kubenswrapper[4578]: I1003 12:55:12.547303 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qdvvk" Oct 03 12:55:12 crc kubenswrapper[4578]: I1003 12:55:12.616237 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k4z8f"] Oct 03 12:55:12 crc kubenswrapper[4578]: I1003 12:55:12.616494 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k4z8f" podUID="445697f9-cf9b-44d0-b76a-c75e6910b758" containerName="registry-server" containerID="cri-o://defd005cc78445af279f2734b9183efb0d9288da3432f00d4abad565e0ac6c99" gracePeriod=2 Oct 03 12:55:12 crc kubenswrapper[4578]: I1003 12:55:12.667714 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-94bd2" podUID="2244329a-bbb1-4c5a-a170-55106dcb21c2" containerName="registry-server" containerID="cri-o://0f7c32f6c75cc7ee3ee180c5115c82b1062f762be4bf0720ab7c446af7f634b3" gracePeriod=2 Oct 03 12:55:12 crc kubenswrapper[4578]: I1003 12:55:12.974556 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k4z8f" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.027659 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-94bd2" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.131859 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2244329a-bbb1-4c5a-a170-55106dcb21c2-catalog-content\") pod \"2244329a-bbb1-4c5a-a170-55106dcb21c2\" (UID: \"2244329a-bbb1-4c5a-a170-55106dcb21c2\") " Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.131942 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55rzg\" (UniqueName: \"kubernetes.io/projected/445697f9-cf9b-44d0-b76a-c75e6910b758-kube-api-access-55rzg\") pod \"445697f9-cf9b-44d0-b76a-c75e6910b758\" (UID: \"445697f9-cf9b-44d0-b76a-c75e6910b758\") " Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.131973 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/445697f9-cf9b-44d0-b76a-c75e6910b758-utilities\") pod \"445697f9-cf9b-44d0-b76a-c75e6910b758\" (UID: \"445697f9-cf9b-44d0-b76a-c75e6910b758\") " Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.132037 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/445697f9-cf9b-44d0-b76a-c75e6910b758-catalog-content\") pod \"445697f9-cf9b-44d0-b76a-c75e6910b758\" (UID: \"445697f9-cf9b-44d0-b76a-c75e6910b758\") " Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.132055 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ws6gc\" (UniqueName: \"kubernetes.io/projected/2244329a-bbb1-4c5a-a170-55106dcb21c2-kube-api-access-ws6gc\") pod \"2244329a-bbb1-4c5a-a170-55106dcb21c2\" (UID: \"2244329a-bbb1-4c5a-a170-55106dcb21c2\") " Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.132102 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2244329a-bbb1-4c5a-a170-55106dcb21c2-utilities\") pod \"2244329a-bbb1-4c5a-a170-55106dcb21c2\" (UID: \"2244329a-bbb1-4c5a-a170-55106dcb21c2\") " Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.133678 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2244329a-bbb1-4c5a-a170-55106dcb21c2-utilities" (OuterVolumeSpecName: "utilities") pod "2244329a-bbb1-4c5a-a170-55106dcb21c2" (UID: "2244329a-bbb1-4c5a-a170-55106dcb21c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.133792 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/445697f9-cf9b-44d0-b76a-c75e6910b758-utilities" (OuterVolumeSpecName: "utilities") pod "445697f9-cf9b-44d0-b76a-c75e6910b758" (UID: "445697f9-cf9b-44d0-b76a-c75e6910b758"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.137584 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2244329a-bbb1-4c5a-a170-55106dcb21c2-kube-api-access-ws6gc" (OuterVolumeSpecName: "kube-api-access-ws6gc") pod "2244329a-bbb1-4c5a-a170-55106dcb21c2" (UID: "2244329a-bbb1-4c5a-a170-55106dcb21c2"). InnerVolumeSpecName "kube-api-access-ws6gc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.138033 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/445697f9-cf9b-44d0-b76a-c75e6910b758-kube-api-access-55rzg" (OuterVolumeSpecName: "kube-api-access-55rzg") pod "445697f9-cf9b-44d0-b76a-c75e6910b758" (UID: "445697f9-cf9b-44d0-b76a-c75e6910b758"). InnerVolumeSpecName "kube-api-access-55rzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.192945 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2244329a-bbb1-4c5a-a170-55106dcb21c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2244329a-bbb1-4c5a-a170-55106dcb21c2" (UID: "2244329a-bbb1-4c5a-a170-55106dcb21c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.194250 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/445697f9-cf9b-44d0-b76a-c75e6910b758-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "445697f9-cf9b-44d0-b76a-c75e6910b758" (UID: "445697f9-cf9b-44d0-b76a-c75e6910b758"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.234820 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ws6gc\" (UniqueName: \"kubernetes.io/projected/2244329a-bbb1-4c5a-a170-55106dcb21c2-kube-api-access-ws6gc\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.235098 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/445697f9-cf9b-44d0-b76a-c75e6910b758-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.235213 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2244329a-bbb1-4c5a-a170-55106dcb21c2-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.235301 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2244329a-bbb1-4c5a-a170-55106dcb21c2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.235381 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55rzg\" (UniqueName: \"kubernetes.io/projected/445697f9-cf9b-44d0-b76a-c75e6910b758-kube-api-access-55rzg\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.235490 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/445697f9-cf9b-44d0-b76a-c75e6910b758-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.675360 4578 generic.go:334] "Generic (PLEG): container finished" podID="445697f9-cf9b-44d0-b76a-c75e6910b758" containerID="defd005cc78445af279f2734b9183efb0d9288da3432f00d4abad565e0ac6c99" exitCode=0 Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.675417 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k4z8f" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.675441 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4z8f" event={"ID":"445697f9-cf9b-44d0-b76a-c75e6910b758","Type":"ContainerDied","Data":"defd005cc78445af279f2734b9183efb0d9288da3432f00d4abad565e0ac6c99"} Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.675472 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k4z8f" event={"ID":"445697f9-cf9b-44d0-b76a-c75e6910b758","Type":"ContainerDied","Data":"431146cd9bcbc4268b4097e1cc64cbc660db407c1e4c3f9cb166bb02b223ac38"} Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.675545 4578 scope.go:117] "RemoveContainer" containerID="defd005cc78445af279f2734b9183efb0d9288da3432f00d4abad565e0ac6c99" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.679495 4578 generic.go:334] "Generic (PLEG): container finished" podID="2244329a-bbb1-4c5a-a170-55106dcb21c2" containerID="0f7c32f6c75cc7ee3ee180c5115c82b1062f762be4bf0720ab7c446af7f634b3" exitCode=0 Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.679535 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-94bd2" event={"ID":"2244329a-bbb1-4c5a-a170-55106dcb21c2","Type":"ContainerDied","Data":"0f7c32f6c75cc7ee3ee180c5115c82b1062f762be4bf0720ab7c446af7f634b3"} Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.679569 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-94bd2" event={"ID":"2244329a-bbb1-4c5a-a170-55106dcb21c2","Type":"ContainerDied","Data":"caee4924e18253ed96aae28aa62a3018f45ed38c5126987554a77a937ad688b3"} Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.679652 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-94bd2" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.694101 4578 scope.go:117] "RemoveContainer" containerID="99d2098662c456703a44270cb86d8f93b1d295f27ffa31a25d5954469f55001d" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.714828 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-94bd2"] Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.718493 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-94bd2"] Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.720594 4578 scope.go:117] "RemoveContainer" containerID="5cb73ff0eae1a1350588654698e7e1e42941c1bfac0342e201d4095cb5622199" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.734799 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k4z8f"] Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.739517 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k4z8f"] Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.754741 4578 scope.go:117] "RemoveContainer" containerID="defd005cc78445af279f2734b9183efb0d9288da3432f00d4abad565e0ac6c99" Oct 03 12:55:13 crc kubenswrapper[4578]: E1003 12:55:13.755293 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"defd005cc78445af279f2734b9183efb0d9288da3432f00d4abad565e0ac6c99\": container with ID starting with defd005cc78445af279f2734b9183efb0d9288da3432f00d4abad565e0ac6c99 not found: ID does not exist" containerID="defd005cc78445af279f2734b9183efb0d9288da3432f00d4abad565e0ac6c99" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.755336 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"defd005cc78445af279f2734b9183efb0d9288da3432f00d4abad565e0ac6c99"} err="failed to get container status \"defd005cc78445af279f2734b9183efb0d9288da3432f00d4abad565e0ac6c99\": rpc error: code = NotFound desc = could not find container \"defd005cc78445af279f2734b9183efb0d9288da3432f00d4abad565e0ac6c99\": container with ID starting with defd005cc78445af279f2734b9183efb0d9288da3432f00d4abad565e0ac6c99 not found: ID does not exist" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.755366 4578 scope.go:117] "RemoveContainer" containerID="99d2098662c456703a44270cb86d8f93b1d295f27ffa31a25d5954469f55001d" Oct 03 12:55:13 crc kubenswrapper[4578]: E1003 12:55:13.755573 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99d2098662c456703a44270cb86d8f93b1d295f27ffa31a25d5954469f55001d\": container with ID starting with 99d2098662c456703a44270cb86d8f93b1d295f27ffa31a25d5954469f55001d not found: ID does not exist" containerID="99d2098662c456703a44270cb86d8f93b1d295f27ffa31a25d5954469f55001d" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.755602 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d2098662c456703a44270cb86d8f93b1d295f27ffa31a25d5954469f55001d"} err="failed to get container status \"99d2098662c456703a44270cb86d8f93b1d295f27ffa31a25d5954469f55001d\": rpc error: code = NotFound desc = could not find container \"99d2098662c456703a44270cb86d8f93b1d295f27ffa31a25d5954469f55001d\": container with ID starting with 99d2098662c456703a44270cb86d8f93b1d295f27ffa31a25d5954469f55001d not found: ID does not exist" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.755620 4578 scope.go:117] "RemoveContainer" containerID="5cb73ff0eae1a1350588654698e7e1e42941c1bfac0342e201d4095cb5622199" Oct 03 12:55:13 crc kubenswrapper[4578]: E1003 12:55:13.755871 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cb73ff0eae1a1350588654698e7e1e42941c1bfac0342e201d4095cb5622199\": container with ID starting with 5cb73ff0eae1a1350588654698e7e1e42941c1bfac0342e201d4095cb5622199 not found: ID does not exist" containerID="5cb73ff0eae1a1350588654698e7e1e42941c1bfac0342e201d4095cb5622199" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.755896 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cb73ff0eae1a1350588654698e7e1e42941c1bfac0342e201d4095cb5622199"} err="failed to get container status \"5cb73ff0eae1a1350588654698e7e1e42941c1bfac0342e201d4095cb5622199\": rpc error: code = NotFound desc = could not find container \"5cb73ff0eae1a1350588654698e7e1e42941c1bfac0342e201d4095cb5622199\": container with ID starting with 5cb73ff0eae1a1350588654698e7e1e42941c1bfac0342e201d4095cb5622199 not found: ID does not exist" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.755913 4578 scope.go:117] "RemoveContainer" containerID="0f7c32f6c75cc7ee3ee180c5115c82b1062f762be4bf0720ab7c446af7f634b3" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.781181 4578 scope.go:117] "RemoveContainer" containerID="7ddf64ec0e9a781add1eaebdcb55208d8b10f4aa18b5ff52613b5dc099126f54" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.793527 4578 scope.go:117] "RemoveContainer" containerID="6b5f2a5de88e4dd4a49eaa028f75af62c13d0a0fd88c6305b603f5cc8674e353" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.820097 4578 scope.go:117] "RemoveContainer" containerID="0f7c32f6c75cc7ee3ee180c5115c82b1062f762be4bf0720ab7c446af7f634b3" Oct 03 12:55:13 crc kubenswrapper[4578]: E1003 12:55:13.821141 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f7c32f6c75cc7ee3ee180c5115c82b1062f762be4bf0720ab7c446af7f634b3\": container with ID starting with 0f7c32f6c75cc7ee3ee180c5115c82b1062f762be4bf0720ab7c446af7f634b3 not found: ID does not exist" containerID="0f7c32f6c75cc7ee3ee180c5115c82b1062f762be4bf0720ab7c446af7f634b3" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.821232 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f7c32f6c75cc7ee3ee180c5115c82b1062f762be4bf0720ab7c446af7f634b3"} err="failed to get container status \"0f7c32f6c75cc7ee3ee180c5115c82b1062f762be4bf0720ab7c446af7f634b3\": rpc error: code = NotFound desc = could not find container \"0f7c32f6c75cc7ee3ee180c5115c82b1062f762be4bf0720ab7c446af7f634b3\": container with ID starting with 0f7c32f6c75cc7ee3ee180c5115c82b1062f762be4bf0720ab7c446af7f634b3 not found: ID does not exist" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.821265 4578 scope.go:117] "RemoveContainer" containerID="7ddf64ec0e9a781add1eaebdcb55208d8b10f4aa18b5ff52613b5dc099126f54" Oct 03 12:55:13 crc kubenswrapper[4578]: E1003 12:55:13.821558 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ddf64ec0e9a781add1eaebdcb55208d8b10f4aa18b5ff52613b5dc099126f54\": container with ID starting with 7ddf64ec0e9a781add1eaebdcb55208d8b10f4aa18b5ff52613b5dc099126f54 not found: ID does not exist" containerID="7ddf64ec0e9a781add1eaebdcb55208d8b10f4aa18b5ff52613b5dc099126f54" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.821581 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ddf64ec0e9a781add1eaebdcb55208d8b10f4aa18b5ff52613b5dc099126f54"} err="failed to get container status \"7ddf64ec0e9a781add1eaebdcb55208d8b10f4aa18b5ff52613b5dc099126f54\": rpc error: code = NotFound desc = could not find container \"7ddf64ec0e9a781add1eaebdcb55208d8b10f4aa18b5ff52613b5dc099126f54\": container with ID starting with 7ddf64ec0e9a781add1eaebdcb55208d8b10f4aa18b5ff52613b5dc099126f54 not found: ID does not exist" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.821597 4578 scope.go:117] "RemoveContainer" containerID="6b5f2a5de88e4dd4a49eaa028f75af62c13d0a0fd88c6305b603f5cc8674e353" Oct 03 12:55:13 crc kubenswrapper[4578]: E1003 12:55:13.821821 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b5f2a5de88e4dd4a49eaa028f75af62c13d0a0fd88c6305b603f5cc8674e353\": container with ID starting with 6b5f2a5de88e4dd4a49eaa028f75af62c13d0a0fd88c6305b603f5cc8674e353 not found: ID does not exist" containerID="6b5f2a5de88e4dd4a49eaa028f75af62c13d0a0fd88c6305b603f5cc8674e353" Oct 03 12:55:13 crc kubenswrapper[4578]: I1003 12:55:13.821845 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b5f2a5de88e4dd4a49eaa028f75af62c13d0a0fd88c6305b603f5cc8674e353"} err="failed to get container status \"6b5f2a5de88e4dd4a49eaa028f75af62c13d0a0fd88c6305b603f5cc8674e353\": rpc error: code = NotFound desc = could not find container \"6b5f2a5de88e4dd4a49eaa028f75af62c13d0a0fd88c6305b603f5cc8674e353\": container with ID starting with 6b5f2a5de88e4dd4a49eaa028f75af62c13d0a0fd88c6305b603f5cc8674e353 not found: ID does not exist" Oct 03 12:55:14 crc kubenswrapper[4578]: I1003 12:55:14.915729 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2244329a-bbb1-4c5a-a170-55106dcb21c2" path="/var/lib/kubelet/pods/2244329a-bbb1-4c5a-a170-55106dcb21c2/volumes" Oct 03 12:55:14 crc kubenswrapper[4578]: I1003 12:55:14.917019 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="445697f9-cf9b-44d0-b76a-c75e6910b758" path="/var/lib/kubelet/pods/445697f9-cf9b-44d0-b76a-c75e6910b758/volumes" Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.013456 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qdvvk"] Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.013738 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qdvvk" podUID="fc33b180-1050-403c-81b7-00ffdf579cd7" containerName="registry-server" containerID="cri-o://f5f0401a8724dfcf17f3d9626ddd580c1e9ee118d0262811f78d8c4842c8ad19" gracePeriod=2 Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.373382 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qdvvk" Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.463014 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmvw5\" (UniqueName: \"kubernetes.io/projected/fc33b180-1050-403c-81b7-00ffdf579cd7-kube-api-access-xmvw5\") pod \"fc33b180-1050-403c-81b7-00ffdf579cd7\" (UID: \"fc33b180-1050-403c-81b7-00ffdf579cd7\") " Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.463171 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc33b180-1050-403c-81b7-00ffdf579cd7-catalog-content\") pod \"fc33b180-1050-403c-81b7-00ffdf579cd7\" (UID: \"fc33b180-1050-403c-81b7-00ffdf579cd7\") " Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.463232 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc33b180-1050-403c-81b7-00ffdf579cd7-utilities\") pod \"fc33b180-1050-403c-81b7-00ffdf579cd7\" (UID: \"fc33b180-1050-403c-81b7-00ffdf579cd7\") " Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.464249 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc33b180-1050-403c-81b7-00ffdf579cd7-utilities" (OuterVolumeSpecName: "utilities") pod "fc33b180-1050-403c-81b7-00ffdf579cd7" (UID: "fc33b180-1050-403c-81b7-00ffdf579cd7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.472168 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc33b180-1050-403c-81b7-00ffdf579cd7-kube-api-access-xmvw5" (OuterVolumeSpecName: "kube-api-access-xmvw5") pod "fc33b180-1050-403c-81b7-00ffdf579cd7" (UID: "fc33b180-1050-403c-81b7-00ffdf579cd7"). InnerVolumeSpecName "kube-api-access-xmvw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.476677 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc33b180-1050-403c-81b7-00ffdf579cd7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc33b180-1050-403c-81b7-00ffdf579cd7" (UID: "fc33b180-1050-403c-81b7-00ffdf579cd7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.564212 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc33b180-1050-403c-81b7-00ffdf579cd7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.564243 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc33b180-1050-403c-81b7-00ffdf579cd7-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.564253 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmvw5\" (UniqueName: \"kubernetes.io/projected/fc33b180-1050-403c-81b7-00ffdf579cd7-kube-api-access-xmvw5\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.691785 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qdvvk" Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.691790 4578 generic.go:334] "Generic (PLEG): container finished" podID="fc33b180-1050-403c-81b7-00ffdf579cd7" containerID="f5f0401a8724dfcf17f3d9626ddd580c1e9ee118d0262811f78d8c4842c8ad19" exitCode=0 Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.691816 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdvvk" event={"ID":"fc33b180-1050-403c-81b7-00ffdf579cd7","Type":"ContainerDied","Data":"f5f0401a8724dfcf17f3d9626ddd580c1e9ee118d0262811f78d8c4842c8ad19"} Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.691858 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdvvk" event={"ID":"fc33b180-1050-403c-81b7-00ffdf579cd7","Type":"ContainerDied","Data":"a85ac3e1addae15d336b28b97794e316043c90bc5df28586a634f67ab69d9253"} Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.691877 4578 scope.go:117] "RemoveContainer" containerID="f5f0401a8724dfcf17f3d9626ddd580c1e9ee118d0262811f78d8c4842c8ad19" Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.707242 4578 scope.go:117] "RemoveContainer" containerID="045dd477c0852a3a82603181c21fe9f46db6f6d09f5f0f54d58c0770bae6cddd" Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.718726 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qdvvk"] Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.721435 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qdvvk"] Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.739130 4578 scope.go:117] "RemoveContainer" containerID="b0d50e8d76c0ad56a7adb095f0099c8062d9026c5db226e4ceb765e2d82d19eb" Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.750223 4578 scope.go:117] "RemoveContainer" containerID="f5f0401a8724dfcf17f3d9626ddd580c1e9ee118d0262811f78d8c4842c8ad19" Oct 03 12:55:15 crc kubenswrapper[4578]: E1003 12:55:15.750568 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5f0401a8724dfcf17f3d9626ddd580c1e9ee118d0262811f78d8c4842c8ad19\": container with ID starting with f5f0401a8724dfcf17f3d9626ddd580c1e9ee118d0262811f78d8c4842c8ad19 not found: ID does not exist" containerID="f5f0401a8724dfcf17f3d9626ddd580c1e9ee118d0262811f78d8c4842c8ad19" Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.750736 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5f0401a8724dfcf17f3d9626ddd580c1e9ee118d0262811f78d8c4842c8ad19"} err="failed to get container status \"f5f0401a8724dfcf17f3d9626ddd580c1e9ee118d0262811f78d8c4842c8ad19\": rpc error: code = NotFound desc = could not find container \"f5f0401a8724dfcf17f3d9626ddd580c1e9ee118d0262811f78d8c4842c8ad19\": container with ID starting with f5f0401a8724dfcf17f3d9626ddd580c1e9ee118d0262811f78d8c4842c8ad19 not found: ID does not exist" Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.750854 4578 scope.go:117] "RemoveContainer" containerID="045dd477c0852a3a82603181c21fe9f46db6f6d09f5f0f54d58c0770bae6cddd" Oct 03 12:55:15 crc kubenswrapper[4578]: E1003 12:55:15.751187 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"045dd477c0852a3a82603181c21fe9f46db6f6d09f5f0f54d58c0770bae6cddd\": container with ID starting with 045dd477c0852a3a82603181c21fe9f46db6f6d09f5f0f54d58c0770bae6cddd not found: ID does not exist" containerID="045dd477c0852a3a82603181c21fe9f46db6f6d09f5f0f54d58c0770bae6cddd" Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.751235 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"045dd477c0852a3a82603181c21fe9f46db6f6d09f5f0f54d58c0770bae6cddd"} err="failed to get container status \"045dd477c0852a3a82603181c21fe9f46db6f6d09f5f0f54d58c0770bae6cddd\": rpc error: code = NotFound desc = could not find container \"045dd477c0852a3a82603181c21fe9f46db6f6d09f5f0f54d58c0770bae6cddd\": container with ID starting with 045dd477c0852a3a82603181c21fe9f46db6f6d09f5f0f54d58c0770bae6cddd not found: ID does not exist" Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.751270 4578 scope.go:117] "RemoveContainer" containerID="b0d50e8d76c0ad56a7adb095f0099c8062d9026c5db226e4ceb765e2d82d19eb" Oct 03 12:55:15 crc kubenswrapper[4578]: E1003 12:55:15.751541 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0d50e8d76c0ad56a7adb095f0099c8062d9026c5db226e4ceb765e2d82d19eb\": container with ID starting with b0d50e8d76c0ad56a7adb095f0099c8062d9026c5db226e4ceb765e2d82d19eb not found: ID does not exist" containerID="b0d50e8d76c0ad56a7adb095f0099c8062d9026c5db226e4ceb765e2d82d19eb" Oct 03 12:55:15 crc kubenswrapper[4578]: I1003 12:55:15.751566 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0d50e8d76c0ad56a7adb095f0099c8062d9026c5db226e4ceb765e2d82d19eb"} err="failed to get container status \"b0d50e8d76c0ad56a7adb095f0099c8062d9026c5db226e4ceb765e2d82d19eb\": rpc error: code = NotFound desc = could not find container \"b0d50e8d76c0ad56a7adb095f0099c8062d9026c5db226e4ceb765e2d82d19eb\": container with ID starting with b0d50e8d76c0ad56a7adb095f0099c8062d9026c5db226e4ceb765e2d82d19eb not found: ID does not exist" Oct 03 12:55:16 crc kubenswrapper[4578]: I1003 12:55:16.916229 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc33b180-1050-403c-81b7-00ffdf579cd7" path="/var/lib/kubelet/pods/fc33b180-1050-403c-81b7-00ffdf579cd7/volumes" Oct 03 12:55:37 crc kubenswrapper[4578]: I1003 12:55:37.288332 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" podUID="89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" containerName="oauth-openshift" containerID="cri-o://ad8012c352db1e465801c2720df816a0a606f653694a1696df01fa45a7494015" gracePeriod=15 Oct 03 12:55:37 crc kubenswrapper[4578]: I1003 12:55:37.803494 4578 generic.go:334] "Generic (PLEG): container finished" podID="89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" containerID="ad8012c352db1e465801c2720df816a0a606f653694a1696df01fa45a7494015" exitCode=0 Oct 03 12:55:37 crc kubenswrapper[4578]: I1003 12:55:37.803572 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" event={"ID":"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e","Type":"ContainerDied","Data":"ad8012c352db1e465801c2720df816a0a606f653694a1696df01fa45a7494015"} Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.185783 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233204 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-657494565c-px4sv"] Oct 03 12:55:38 crc kubenswrapper[4578]: E1003 12:55:38.233479 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2244329a-bbb1-4c5a-a170-55106dcb21c2" containerName="extract-utilities" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233499 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="2244329a-bbb1-4c5a-a170-55106dcb21c2" containerName="extract-utilities" Oct 03 12:55:38 crc kubenswrapper[4578]: E1003 12:55:38.233514 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2244329a-bbb1-4c5a-a170-55106dcb21c2" containerName="registry-server" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233522 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="2244329a-bbb1-4c5a-a170-55106dcb21c2" containerName="registry-server" Oct 03 12:55:38 crc kubenswrapper[4578]: E1003 12:55:38.233533 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" containerName="oauth-openshift" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233541 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" containerName="oauth-openshift" Oct 03 12:55:38 crc kubenswrapper[4578]: E1003 12:55:38.233552 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc33b180-1050-403c-81b7-00ffdf579cd7" containerName="extract-content" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233560 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc33b180-1050-403c-81b7-00ffdf579cd7" containerName="extract-content" Oct 03 12:55:38 crc kubenswrapper[4578]: E1003 12:55:38.233571 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4754126b-ff64-44ce-829d-6f8307eed3a7" containerName="registry-server" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233580 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="4754126b-ff64-44ce-829d-6f8307eed3a7" containerName="registry-server" Oct 03 12:55:38 crc kubenswrapper[4578]: E1003 12:55:38.233590 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4754126b-ff64-44ce-829d-6f8307eed3a7" containerName="extract-utilities" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233597 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="4754126b-ff64-44ce-829d-6f8307eed3a7" containerName="extract-utilities" Oct 03 12:55:38 crc kubenswrapper[4578]: E1003 12:55:38.233609 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="445697f9-cf9b-44d0-b76a-c75e6910b758" containerName="registry-server" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233616 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="445697f9-cf9b-44d0-b76a-c75e6910b758" containerName="registry-server" Oct 03 12:55:38 crc kubenswrapper[4578]: E1003 12:55:38.233629 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2242cf82-497d-4e44-bacb-0da986ee2dcb" containerName="pruner" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233647 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="2242cf82-497d-4e44-bacb-0da986ee2dcb" containerName="pruner" Oct 03 12:55:38 crc kubenswrapper[4578]: E1003 12:55:38.233659 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4754126b-ff64-44ce-829d-6f8307eed3a7" containerName="extract-content" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233667 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="4754126b-ff64-44ce-829d-6f8307eed3a7" containerName="extract-content" Oct 03 12:55:38 crc kubenswrapper[4578]: E1003 12:55:38.233680 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="445697f9-cf9b-44d0-b76a-c75e6910b758" containerName="extract-utilities" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233687 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="445697f9-cf9b-44d0-b76a-c75e6910b758" containerName="extract-utilities" Oct 03 12:55:38 crc kubenswrapper[4578]: E1003 12:55:38.233701 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2244329a-bbb1-4c5a-a170-55106dcb21c2" containerName="extract-content" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233709 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="2244329a-bbb1-4c5a-a170-55106dcb21c2" containerName="extract-content" Oct 03 12:55:38 crc kubenswrapper[4578]: E1003 12:55:38.233718 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc33b180-1050-403c-81b7-00ffdf579cd7" containerName="registry-server" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233725 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc33b180-1050-403c-81b7-00ffdf579cd7" containerName="registry-server" Oct 03 12:55:38 crc kubenswrapper[4578]: E1003 12:55:38.233737 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdfd4313-d05e-4cbd-8520-3f929675c662" containerName="collect-profiles" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233746 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdfd4313-d05e-4cbd-8520-3f929675c662" containerName="collect-profiles" Oct 03 12:55:38 crc kubenswrapper[4578]: E1003 12:55:38.233757 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30c6876c-90e4-4670-999d-64556e223478" containerName="pruner" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233765 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="30c6876c-90e4-4670-999d-64556e223478" containerName="pruner" Oct 03 12:55:38 crc kubenswrapper[4578]: E1003 12:55:38.233777 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="445697f9-cf9b-44d0-b76a-c75e6910b758" containerName="extract-content" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233783 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="445697f9-cf9b-44d0-b76a-c75e6910b758" containerName="extract-content" Oct 03 12:55:38 crc kubenswrapper[4578]: E1003 12:55:38.233792 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc33b180-1050-403c-81b7-00ffdf579cd7" containerName="extract-utilities" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233798 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc33b180-1050-403c-81b7-00ffdf579cd7" containerName="extract-utilities" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233905 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="2242cf82-497d-4e44-bacb-0da986ee2dcb" containerName="pruner" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233917 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc33b180-1050-403c-81b7-00ffdf579cd7" containerName="registry-server" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233927 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="2244329a-bbb1-4c5a-a170-55106dcb21c2" containerName="registry-server" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233938 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" containerName="oauth-openshift" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233950 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="445697f9-cf9b-44d0-b76a-c75e6910b758" containerName="registry-server" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233959 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="4754126b-ff64-44ce-829d-6f8307eed3a7" containerName="registry-server" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233968 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="30c6876c-90e4-4670-999d-64556e223478" containerName="pruner" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.233982 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdfd4313-d05e-4cbd-8520-3f929675c662" containerName="collect-profiles" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.234380 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.239470 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-657494565c-px4sv"] Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.298645 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkv6n\" (UniqueName: \"kubernetes.io/projected/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-kube-api-access-nkv6n\") pod \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.298700 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-template-login\") pod \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.298730 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-template-provider-selection\") pod \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.298861 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-session\") pod \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.298900 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-ocp-branding-template\") pod \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.298931 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-router-certs\") pod \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.298955 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-audit-policies\") pod \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.298978 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-template-error\") pod \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.299009 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-audit-dir\") pod \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.299041 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-service-ca\") pod \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.299068 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-trusted-ca-bundle\") pod \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.299103 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-idp-0-file-data\") pod \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.299131 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-cliconfig\") pod \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.299157 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-serving-cert\") pod \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\" (UID: \"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e\") " Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.300023 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" (UID: "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.300955 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" (UID: "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.304731 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" (UID: "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.305176 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" (UID: "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.305253 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" (UID: "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.305527 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" (UID: "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.305907 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" (UID: "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.305999 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" (UID: "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.306034 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-kube-api-access-nkv6n" (OuterVolumeSpecName: "kube-api-access-nkv6n") pod "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" (UID: "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e"). InnerVolumeSpecName "kube-api-access-nkv6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.312831 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" (UID: "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.313276 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" (UID: "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.313695 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" (UID: "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.315154 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" (UID: "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.315694 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" (UID: "89b4a9e4-3b97-4ebc-850e-8995efbbaf7e"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.400866 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.400947 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-user-template-error\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.400990 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.401019 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2rsp\" (UniqueName: \"kubernetes.io/projected/282727fe-9818-482d-be67-e28274cf7ed6-kube-api-access-r2rsp\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.401102 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-user-template-login\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.401444 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.401496 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/282727fe-9818-482d-be67-e28274cf7ed6-audit-policies\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.401513 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.401534 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.401554 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-router-certs\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.401864 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-session\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.401948 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.401981 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/282727fe-9818-482d-be67-e28274cf7ed6-audit-dir\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.402005 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-service-ca\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.402074 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.402092 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.402102 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkv6n\" (UniqueName: \"kubernetes.io/projected/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-kube-api-access-nkv6n\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.402113 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.402123 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.402133 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.402142 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.402151 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.402162 4578 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.402171 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.402181 4578 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.402190 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.402199 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.402210 4578 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.503325 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-session\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.503407 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.503451 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/282727fe-9818-482d-be67-e28274cf7ed6-audit-dir\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.503480 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-service-ca\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.503523 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.503565 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-user-template-error\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.503597 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.503633 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2rsp\" (UniqueName: \"kubernetes.io/projected/282727fe-9818-482d-be67-e28274cf7ed6-kube-api-access-r2rsp\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.503675 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/282727fe-9818-482d-be67-e28274cf7ed6-audit-dir\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.503720 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-user-template-login\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.503761 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.503804 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/282727fe-9818-482d-be67-e28274cf7ed6-audit-policies\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.503840 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.503873 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.503909 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-router-certs\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.509351 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-router-certs\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.509638 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.510855 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.510864 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.510976 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.511353 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.512149 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.515237 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-user-template-login\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.515587 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-service-ca\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.515874 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-system-session\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.518177 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/282727fe-9818-482d-be67-e28274cf7ed6-audit-policies\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.518224 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/282727fe-9818-482d-be67-e28274cf7ed6-v4-0-config-user-template-error\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.532789 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2rsp\" (UniqueName: \"kubernetes.io/projected/282727fe-9818-482d-be67-e28274cf7ed6-kube-api-access-r2rsp\") pod \"oauth-openshift-657494565c-px4sv\" (UID: \"282727fe-9818-482d-be67-e28274cf7ed6\") " pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.550690 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.810355 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" event={"ID":"89b4a9e4-3b97-4ebc-850e-8995efbbaf7e","Type":"ContainerDied","Data":"6d3abbb8b3f38e30706c3b831a4b9960371b460deb29c40ac145ecc356e21e58"} Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.810750 4578 scope.go:117] "RemoveContainer" containerID="ad8012c352db1e465801c2720df816a0a606f653694a1696df01fa45a7494015" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.810493 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7br6x" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.837544 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7br6x"] Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.840350 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7br6x"] Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.919256 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89b4a9e4-3b97-4ebc-850e-8995efbbaf7e" path="/var/lib/kubelet/pods/89b4a9e4-3b97-4ebc-850e-8995efbbaf7e/volumes" Oct 03 12:55:38 crc kubenswrapper[4578]: I1003 12:55:38.929936 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-657494565c-px4sv"] Oct 03 12:55:39 crc kubenswrapper[4578]: I1003 12:55:39.817490 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-657494565c-px4sv" event={"ID":"282727fe-9818-482d-be67-e28274cf7ed6","Type":"ContainerStarted","Data":"5ae01f2134a78006db969596bbe382c2a995c9fb8538f1bf422afcbef42a3462"} Oct 03 12:55:39 crc kubenswrapper[4578]: I1003 12:55:39.817828 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:39 crc kubenswrapper[4578]: I1003 12:55:39.817844 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-657494565c-px4sv" event={"ID":"282727fe-9818-482d-be67-e28274cf7ed6","Type":"ContainerStarted","Data":"8e10805f0a61c38f754bfd786b12eac0d0807855b000d1cfb77b0e6388760fb9"} Oct 03 12:55:39 crc kubenswrapper[4578]: I1003 12:55:39.822999 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-657494565c-px4sv" Oct 03 12:55:39 crc kubenswrapper[4578]: I1003 12:55:39.837404 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-657494565c-px4sv" podStartSLOduration=27.837383888 podStartE2EDuration="27.837383888s" podCreationTimestamp="2025-10-03 12:55:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:55:39.835245117 +0000 UTC m=+275.633717311" watchObservedRunningTime="2025-10-03 12:55:39.837383888 +0000 UTC m=+275.635856082" Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.385366 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hmk8t"] Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.386166 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hmk8t" podUID="742a4571-0a86-4cb4-bb93-8b46219f6d49" containerName="registry-server" containerID="cri-o://42b16065e5cd5959e35719370e31b905f5279cd890408d7c5543fa49e060d5a5" gracePeriod=30 Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.398264 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ldfq4"] Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.399008 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ldfq4" podUID="b40b7869-c4f0-427d-b99c-fca7f0d8b937" containerName="registry-server" containerID="cri-o://0eb7978f73b5b71a41f58d92e0989daa504e621cc70de06a65d413e9353b45bb" gracePeriod=30 Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.413650 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sqbkd"] Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.413901 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" podUID="4ef05574-4924-4a73-b0ba-0e632c96410a" containerName="marketplace-operator" containerID="cri-o://439daae67e164ac422ee4754f63691531024d81e693b41b096aecd99ecedc43d" gracePeriod=30 Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.418862 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5hjv"] Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.419351 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w5hjv" podUID="75b24dc1-77ec-4b54-9f85-2443e273eadf" containerName="registry-server" containerID="cri-o://c0443a230b0f1a59a152fd9fe5e5afab476c4c195bdbe35c0a33446f5e491377" gracePeriod=30 Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.426083 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n6875"] Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.426304 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n6875" podUID="beb5b22e-cc9e-453b-bac2-88f7a5d30c29" containerName="registry-server" containerID="cri-o://d6dfbd083ed07317b9c09ccde3afa3a6e322e7bab844def4d21e504843685d41" gracePeriod=30 Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.438745 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bf7p9"] Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.439337 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bf7p9" Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.454564 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bf7p9"] Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.574673 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k82lf\" (UniqueName: \"kubernetes.io/projected/30b79cea-e428-4de1-897c-f561c1bfc659-kube-api-access-k82lf\") pod \"marketplace-operator-79b997595-bf7p9\" (UID: \"30b79cea-e428-4de1-897c-f561c1bfc659\") " pod="openshift-marketplace/marketplace-operator-79b997595-bf7p9" Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.574750 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/30b79cea-e428-4de1-897c-f561c1bfc659-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bf7p9\" (UID: \"30b79cea-e428-4de1-897c-f561c1bfc659\") " pod="openshift-marketplace/marketplace-operator-79b997595-bf7p9" Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.574772 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/30b79cea-e428-4de1-897c-f561c1bfc659-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bf7p9\" (UID: \"30b79cea-e428-4de1-897c-f561c1bfc659\") " pod="openshift-marketplace/marketplace-operator-79b997595-bf7p9" Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.695481 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/30b79cea-e428-4de1-897c-f561c1bfc659-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bf7p9\" (UID: \"30b79cea-e428-4de1-897c-f561c1bfc659\") " pod="openshift-marketplace/marketplace-operator-79b997595-bf7p9" Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.695552 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/30b79cea-e428-4de1-897c-f561c1bfc659-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bf7p9\" (UID: \"30b79cea-e428-4de1-897c-f561c1bfc659\") " pod="openshift-marketplace/marketplace-operator-79b997595-bf7p9" Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.696986 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/30b79cea-e428-4de1-897c-f561c1bfc659-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bf7p9\" (UID: \"30b79cea-e428-4de1-897c-f561c1bfc659\") " pod="openshift-marketplace/marketplace-operator-79b997595-bf7p9" Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.697048 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k82lf\" (UniqueName: \"kubernetes.io/projected/30b79cea-e428-4de1-897c-f561c1bfc659-kube-api-access-k82lf\") pod \"marketplace-operator-79b997595-bf7p9\" (UID: \"30b79cea-e428-4de1-897c-f561c1bfc659\") " pod="openshift-marketplace/marketplace-operator-79b997595-bf7p9" Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.722713 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k82lf\" (UniqueName: \"kubernetes.io/projected/30b79cea-e428-4de1-897c-f561c1bfc659-kube-api-access-k82lf\") pod \"marketplace-operator-79b997595-bf7p9\" (UID: \"30b79cea-e428-4de1-897c-f561c1bfc659\") " pod="openshift-marketplace/marketplace-operator-79b997595-bf7p9" Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.723248 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/30b79cea-e428-4de1-897c-f561c1bfc659-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bf7p9\" (UID: \"30b79cea-e428-4de1-897c-f561c1bfc659\") " pod="openshift-marketplace/marketplace-operator-79b997595-bf7p9" Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.765993 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bf7p9" Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.804394 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hmk8t" Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.893366 4578 generic.go:334] "Generic (PLEG): container finished" podID="4ef05574-4924-4a73-b0ba-0e632c96410a" containerID="439daae67e164ac422ee4754f63691531024d81e693b41b096aecd99ecedc43d" exitCode=0 Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.893477 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" event={"ID":"4ef05574-4924-4a73-b0ba-0e632c96410a","Type":"ContainerDied","Data":"439daae67e164ac422ee4754f63691531024d81e693b41b096aecd99ecedc43d"} Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.900014 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/742a4571-0a86-4cb4-bb93-8b46219f6d49-catalog-content\") pod \"742a4571-0a86-4cb4-bb93-8b46219f6d49\" (UID: \"742a4571-0a86-4cb4-bb93-8b46219f6d49\") " Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.900064 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rccr\" (UniqueName: \"kubernetes.io/projected/742a4571-0a86-4cb4-bb93-8b46219f6d49-kube-api-access-7rccr\") pod \"742a4571-0a86-4cb4-bb93-8b46219f6d49\" (UID: \"742a4571-0a86-4cb4-bb93-8b46219f6d49\") " Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.900102 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/742a4571-0a86-4cb4-bb93-8b46219f6d49-utilities\") pod \"742a4571-0a86-4cb4-bb93-8b46219f6d49\" (UID: \"742a4571-0a86-4cb4-bb93-8b46219f6d49\") " Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.901585 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/742a4571-0a86-4cb4-bb93-8b46219f6d49-utilities" (OuterVolumeSpecName: "utilities") pod "742a4571-0a86-4cb4-bb93-8b46219f6d49" (UID: "742a4571-0a86-4cb4-bb93-8b46219f6d49"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.902015 4578 generic.go:334] "Generic (PLEG): container finished" podID="beb5b22e-cc9e-453b-bac2-88f7a5d30c29" containerID="d6dfbd083ed07317b9c09ccde3afa3a6e322e7bab844def4d21e504843685d41" exitCode=0 Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.902249 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6875" event={"ID":"beb5b22e-cc9e-453b-bac2-88f7a5d30c29","Type":"ContainerDied","Data":"d6dfbd083ed07317b9c09ccde3afa3a6e322e7bab844def4d21e504843685d41"} Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.909181 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/742a4571-0a86-4cb4-bb93-8b46219f6d49-kube-api-access-7rccr" (OuterVolumeSpecName: "kube-api-access-7rccr") pod "742a4571-0a86-4cb4-bb93-8b46219f6d49" (UID: "742a4571-0a86-4cb4-bb93-8b46219f6d49"). InnerVolumeSpecName "kube-api-access-7rccr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.909771 4578 generic.go:334] "Generic (PLEG): container finished" podID="742a4571-0a86-4cb4-bb93-8b46219f6d49" containerID="42b16065e5cd5959e35719370e31b905f5279cd890408d7c5543fa49e060d5a5" exitCode=0 Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.909836 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hmk8t" event={"ID":"742a4571-0a86-4cb4-bb93-8b46219f6d49","Type":"ContainerDied","Data":"42b16065e5cd5959e35719370e31b905f5279cd890408d7c5543fa49e060d5a5"} Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.909861 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hmk8t" event={"ID":"742a4571-0a86-4cb4-bb93-8b46219f6d49","Type":"ContainerDied","Data":"84225a706ab101fc2aa255d82534aea42e72fc84811bc366ef8b2ff78c78372c"} Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.909881 4578 scope.go:117] "RemoveContainer" containerID="42b16065e5cd5959e35719370e31b905f5279cd890408d7c5543fa49e060d5a5" Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.910015 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hmk8t" Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.926613 4578 generic.go:334] "Generic (PLEG): container finished" podID="b40b7869-c4f0-427d-b99c-fca7f0d8b937" containerID="0eb7978f73b5b71a41f58d92e0989daa504e621cc70de06a65d413e9353b45bb" exitCode=0 Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.926809 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ldfq4" event={"ID":"b40b7869-c4f0-427d-b99c-fca7f0d8b937","Type":"ContainerDied","Data":"0eb7978f73b5b71a41f58d92e0989daa504e621cc70de06a65d413e9353b45bb"} Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.937169 4578 generic.go:334] "Generic (PLEG): container finished" podID="75b24dc1-77ec-4b54-9f85-2443e273eadf" containerID="c0443a230b0f1a59a152fd9fe5e5afab476c4c195bdbe35c0a33446f5e491377" exitCode=0 Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.937232 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5hjv" event={"ID":"75b24dc1-77ec-4b54-9f85-2443e273eadf","Type":"ContainerDied","Data":"c0443a230b0f1a59a152fd9fe5e5afab476c4c195bdbe35c0a33446f5e491377"} Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.938587 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" Oct 03 12:55:51 crc kubenswrapper[4578]: I1003 12:55:51.959310 4578 scope.go:117] "RemoveContainer" containerID="35e099489bdb34a5f63b43d43b657ee4608e45d458c23deab329cc34f1f8fde0" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.000798 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4ef05574-4924-4a73-b0ba-0e632c96410a-marketplace-trusted-ca\") pod \"4ef05574-4924-4a73-b0ba-0e632c96410a\" (UID: \"4ef05574-4924-4a73-b0ba-0e632c96410a\") " Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.000919 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7m4lf\" (UniqueName: \"kubernetes.io/projected/4ef05574-4924-4a73-b0ba-0e632c96410a-kube-api-access-7m4lf\") pod \"4ef05574-4924-4a73-b0ba-0e632c96410a\" (UID: \"4ef05574-4924-4a73-b0ba-0e632c96410a\") " Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.000943 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4ef05574-4924-4a73-b0ba-0e632c96410a-marketplace-operator-metrics\") pod \"4ef05574-4924-4a73-b0ba-0e632c96410a\" (UID: \"4ef05574-4924-4a73-b0ba-0e632c96410a\") " Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.001153 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rccr\" (UniqueName: \"kubernetes.io/projected/742a4571-0a86-4cb4-bb93-8b46219f6d49-kube-api-access-7rccr\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.001166 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/742a4571-0a86-4cb4-bb93-8b46219f6d49-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.013414 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ef05574-4924-4a73-b0ba-0e632c96410a-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "4ef05574-4924-4a73-b0ba-0e632c96410a" (UID: "4ef05574-4924-4a73-b0ba-0e632c96410a"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.023459 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/742a4571-0a86-4cb4-bb93-8b46219f6d49-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "742a4571-0a86-4cb4-bb93-8b46219f6d49" (UID: "742a4571-0a86-4cb4-bb93-8b46219f6d49"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.044764 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ef05574-4924-4a73-b0ba-0e632c96410a-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "4ef05574-4924-4a73-b0ba-0e632c96410a" (UID: "4ef05574-4924-4a73-b0ba-0e632c96410a"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.045773 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ef05574-4924-4a73-b0ba-0e632c96410a-kube-api-access-7m4lf" (OuterVolumeSpecName: "kube-api-access-7m4lf") pod "4ef05574-4924-4a73-b0ba-0e632c96410a" (UID: "4ef05574-4924-4a73-b0ba-0e632c96410a"). InnerVolumeSpecName "kube-api-access-7m4lf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.083315 4578 scope.go:117] "RemoveContainer" containerID="5a8cdf1cca6e3eef1b5bafa3c1518bf4e4b7a593437be92d88c7fa1df46e83b7" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.107770 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n6875" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.107810 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w5hjv" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.109825 4578 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4ef05574-4924-4a73-b0ba-0e632c96410a-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.109872 4578 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4ef05574-4924-4a73-b0ba-0e632c96410a-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.109898 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/742a4571-0a86-4cb4-bb93-8b46219f6d49-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.109924 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7m4lf\" (UniqueName: \"kubernetes.io/projected/4ef05574-4924-4a73-b0ba-0e632c96410a-kube-api-access-7m4lf\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.121100 4578 scope.go:117] "RemoveContainer" containerID="42b16065e5cd5959e35719370e31b905f5279cd890408d7c5543fa49e060d5a5" Oct 03 12:55:52 crc kubenswrapper[4578]: E1003 12:55:52.122192 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42b16065e5cd5959e35719370e31b905f5279cd890408d7c5543fa49e060d5a5\": container with ID starting with 42b16065e5cd5959e35719370e31b905f5279cd890408d7c5543fa49e060d5a5 not found: ID does not exist" containerID="42b16065e5cd5959e35719370e31b905f5279cd890408d7c5543fa49e060d5a5" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.122221 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42b16065e5cd5959e35719370e31b905f5279cd890408d7c5543fa49e060d5a5"} err="failed to get container status \"42b16065e5cd5959e35719370e31b905f5279cd890408d7c5543fa49e060d5a5\": rpc error: code = NotFound desc = could not find container \"42b16065e5cd5959e35719370e31b905f5279cd890408d7c5543fa49e060d5a5\": container with ID starting with 42b16065e5cd5959e35719370e31b905f5279cd890408d7c5543fa49e060d5a5 not found: ID does not exist" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.122240 4578 scope.go:117] "RemoveContainer" containerID="35e099489bdb34a5f63b43d43b657ee4608e45d458c23deab329cc34f1f8fde0" Oct 03 12:55:52 crc kubenswrapper[4578]: E1003 12:55:52.123462 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35e099489bdb34a5f63b43d43b657ee4608e45d458c23deab329cc34f1f8fde0\": container with ID starting with 35e099489bdb34a5f63b43d43b657ee4608e45d458c23deab329cc34f1f8fde0 not found: ID does not exist" containerID="35e099489bdb34a5f63b43d43b657ee4608e45d458c23deab329cc34f1f8fde0" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.123487 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35e099489bdb34a5f63b43d43b657ee4608e45d458c23deab329cc34f1f8fde0"} err="failed to get container status \"35e099489bdb34a5f63b43d43b657ee4608e45d458c23deab329cc34f1f8fde0\": rpc error: code = NotFound desc = could not find container \"35e099489bdb34a5f63b43d43b657ee4608e45d458c23deab329cc34f1f8fde0\": container with ID starting with 35e099489bdb34a5f63b43d43b657ee4608e45d458c23deab329cc34f1f8fde0 not found: ID does not exist" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.123504 4578 scope.go:117] "RemoveContainer" containerID="5a8cdf1cca6e3eef1b5bafa3c1518bf4e4b7a593437be92d88c7fa1df46e83b7" Oct 03 12:55:52 crc kubenswrapper[4578]: E1003 12:55:52.123901 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a8cdf1cca6e3eef1b5bafa3c1518bf4e4b7a593437be92d88c7fa1df46e83b7\": container with ID starting with 5a8cdf1cca6e3eef1b5bafa3c1518bf4e4b7a593437be92d88c7fa1df46e83b7 not found: ID does not exist" containerID="5a8cdf1cca6e3eef1b5bafa3c1518bf4e4b7a593437be92d88c7fa1df46e83b7" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.123935 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a8cdf1cca6e3eef1b5bafa3c1518bf4e4b7a593437be92d88c7fa1df46e83b7"} err="failed to get container status \"5a8cdf1cca6e3eef1b5bafa3c1518bf4e4b7a593437be92d88c7fa1df46e83b7\": rpc error: code = NotFound desc = could not find container \"5a8cdf1cca6e3eef1b5bafa3c1518bf4e4b7a593437be92d88c7fa1df46e83b7\": container with ID starting with 5a8cdf1cca6e3eef1b5bafa3c1518bf4e4b7a593437be92d88c7fa1df46e83b7 not found: ID does not exist" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.213559 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beb5b22e-cc9e-453b-bac2-88f7a5d30c29-utilities\") pod \"beb5b22e-cc9e-453b-bac2-88f7a5d30c29\" (UID: \"beb5b22e-cc9e-453b-bac2-88f7a5d30c29\") " Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.213747 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75b24dc1-77ec-4b54-9f85-2443e273eadf-utilities\") pod \"75b24dc1-77ec-4b54-9f85-2443e273eadf\" (UID: \"75b24dc1-77ec-4b54-9f85-2443e273eadf\") " Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.213790 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfh44\" (UniqueName: \"kubernetes.io/projected/beb5b22e-cc9e-453b-bac2-88f7a5d30c29-kube-api-access-rfh44\") pod \"beb5b22e-cc9e-453b-bac2-88f7a5d30c29\" (UID: \"beb5b22e-cc9e-453b-bac2-88f7a5d30c29\") " Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.213814 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5bqk\" (UniqueName: \"kubernetes.io/projected/75b24dc1-77ec-4b54-9f85-2443e273eadf-kube-api-access-s5bqk\") pod \"75b24dc1-77ec-4b54-9f85-2443e273eadf\" (UID: \"75b24dc1-77ec-4b54-9f85-2443e273eadf\") " Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.213844 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75b24dc1-77ec-4b54-9f85-2443e273eadf-catalog-content\") pod \"75b24dc1-77ec-4b54-9f85-2443e273eadf\" (UID: \"75b24dc1-77ec-4b54-9f85-2443e273eadf\") " Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.213865 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beb5b22e-cc9e-453b-bac2-88f7a5d30c29-catalog-content\") pod \"beb5b22e-cc9e-453b-bac2-88f7a5d30c29\" (UID: \"beb5b22e-cc9e-453b-bac2-88f7a5d30c29\") " Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.217153 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75b24dc1-77ec-4b54-9f85-2443e273eadf-utilities" (OuterVolumeSpecName: "utilities") pod "75b24dc1-77ec-4b54-9f85-2443e273eadf" (UID: "75b24dc1-77ec-4b54-9f85-2443e273eadf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.224390 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75b24dc1-77ec-4b54-9f85-2443e273eadf-kube-api-access-s5bqk" (OuterVolumeSpecName: "kube-api-access-s5bqk") pod "75b24dc1-77ec-4b54-9f85-2443e273eadf" (UID: "75b24dc1-77ec-4b54-9f85-2443e273eadf"). InnerVolumeSpecName "kube-api-access-s5bqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.229945 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/beb5b22e-cc9e-453b-bac2-88f7a5d30c29-utilities" (OuterVolumeSpecName: "utilities") pod "beb5b22e-cc9e-453b-bac2-88f7a5d30c29" (UID: "beb5b22e-cc9e-453b-bac2-88f7a5d30c29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.230851 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beb5b22e-cc9e-453b-bac2-88f7a5d30c29-kube-api-access-rfh44" (OuterVolumeSpecName: "kube-api-access-rfh44") pod "beb5b22e-cc9e-453b-bac2-88f7a5d30c29" (UID: "beb5b22e-cc9e-453b-bac2-88f7a5d30c29"). InnerVolumeSpecName "kube-api-access-rfh44". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.245911 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75b24dc1-77ec-4b54-9f85-2443e273eadf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "75b24dc1-77ec-4b54-9f85-2443e273eadf" (UID: "75b24dc1-77ec-4b54-9f85-2443e273eadf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.258305 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hmk8t"] Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.262560 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hmk8t"] Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.314533 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75b24dc1-77ec-4b54-9f85-2443e273eadf-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.314561 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfh44\" (UniqueName: \"kubernetes.io/projected/beb5b22e-cc9e-453b-bac2-88f7a5d30c29-kube-api-access-rfh44\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.314576 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5bqk\" (UniqueName: \"kubernetes.io/projected/75b24dc1-77ec-4b54-9f85-2443e273eadf-kube-api-access-s5bqk\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.314587 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75b24dc1-77ec-4b54-9f85-2443e273eadf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.314598 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beb5b22e-cc9e-453b-bac2-88f7a5d30c29-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.329900 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bf7p9"] Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.333061 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/beb5b22e-cc9e-453b-bac2-88f7a5d30c29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "beb5b22e-cc9e-453b-bac2-88f7a5d30c29" (UID: "beb5b22e-cc9e-453b-bac2-88f7a5d30c29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:55:52 crc kubenswrapper[4578]: W1003 12:55:52.345088 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30b79cea_e428_4de1_897c_f561c1bfc659.slice/crio-7e2e8b8669bfc695c25c9f96131a15d6ee5dd5688689f05a68b55246e30f8b3b WatchSource:0}: Error finding container 7e2e8b8669bfc695c25c9f96131a15d6ee5dd5688689f05a68b55246e30f8b3b: Status 404 returned error can't find the container with id 7e2e8b8669bfc695c25c9f96131a15d6ee5dd5688689f05a68b55246e30f8b3b Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.397958 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ldfq4" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.415248 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t287h\" (UniqueName: \"kubernetes.io/projected/b40b7869-c4f0-427d-b99c-fca7f0d8b937-kube-api-access-t287h\") pod \"b40b7869-c4f0-427d-b99c-fca7f0d8b937\" (UID: \"b40b7869-c4f0-427d-b99c-fca7f0d8b937\") " Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.416317 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b40b7869-c4f0-427d-b99c-fca7f0d8b937-utilities\") pod \"b40b7869-c4f0-427d-b99c-fca7f0d8b937\" (UID: \"b40b7869-c4f0-427d-b99c-fca7f0d8b937\") " Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.416367 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b40b7869-c4f0-427d-b99c-fca7f0d8b937-catalog-content\") pod \"b40b7869-c4f0-427d-b99c-fca7f0d8b937\" (UID: \"b40b7869-c4f0-427d-b99c-fca7f0d8b937\") " Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.420351 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b40b7869-c4f0-427d-b99c-fca7f0d8b937-utilities" (OuterVolumeSpecName: "utilities") pod "b40b7869-c4f0-427d-b99c-fca7f0d8b937" (UID: "b40b7869-c4f0-427d-b99c-fca7f0d8b937"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.420595 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beb5b22e-cc9e-453b-bac2-88f7a5d30c29-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.420613 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b40b7869-c4f0-427d-b99c-fca7f0d8b937-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.429838 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b40b7869-c4f0-427d-b99c-fca7f0d8b937-kube-api-access-t287h" (OuterVolumeSpecName: "kube-api-access-t287h") pod "b40b7869-c4f0-427d-b99c-fca7f0d8b937" (UID: "b40b7869-c4f0-427d-b99c-fca7f0d8b937"). InnerVolumeSpecName "kube-api-access-t287h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.472885 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b40b7869-c4f0-427d-b99c-fca7f0d8b937-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b40b7869-c4f0-427d-b99c-fca7f0d8b937" (UID: "b40b7869-c4f0-427d-b99c-fca7f0d8b937"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.520992 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b40b7869-c4f0-427d-b99c-fca7f0d8b937-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.521033 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t287h\" (UniqueName: \"kubernetes.io/projected/b40b7869-c4f0-427d-b99c-fca7f0d8b937-kube-api-access-t287h\") on node \"crc\" DevicePath \"\"" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.915787 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="742a4571-0a86-4cb4-bb93-8b46219f6d49" path="/var/lib/kubelet/pods/742a4571-0a86-4cb4-bb93-8b46219f6d49/volumes" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.946843 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6875" event={"ID":"beb5b22e-cc9e-453b-bac2-88f7a5d30c29","Type":"ContainerDied","Data":"0cb90b217bc38bb23675b8a5d1974f78860e2786216fe338137591b042899772"} Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.946877 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n6875" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.946898 4578 scope.go:117] "RemoveContainer" containerID="d6dfbd083ed07317b9c09ccde3afa3a6e322e7bab844def4d21e504843685d41" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.951936 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bf7p9" event={"ID":"30b79cea-e428-4de1-897c-f561c1bfc659","Type":"ContainerStarted","Data":"3973f221e5675e554733e32af869d39e7338f6cfe07c551595140d6692cb8b32"} Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.951969 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bf7p9" event={"ID":"30b79cea-e428-4de1-897c-f561c1bfc659","Type":"ContainerStarted","Data":"7e2e8b8669bfc695c25c9f96131a15d6ee5dd5688689f05a68b55246e30f8b3b"} Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.952261 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-bf7p9" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.953917 4578 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-bf7p9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.953954 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-bf7p9" podUID="30b79cea-e428-4de1-897c-f561c1bfc659" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.963371 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ldfq4" event={"ID":"b40b7869-c4f0-427d-b99c-fca7f0d8b937","Type":"ContainerDied","Data":"2fc756e984b124eb0a6d4c5d26df07dda09d9cbb1668492b67d8fb1fc589b67a"} Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.963645 4578 scope.go:117] "RemoveContainer" containerID="c5a5f2cbc2c1b9dbbf9f69af41dc4e19a2d2abae656b50a51f469220375645f2" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.964210 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ldfq4" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.969897 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w5hjv" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.970123 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5hjv" event={"ID":"75b24dc1-77ec-4b54-9f85-2443e273eadf","Type":"ContainerDied","Data":"cc9b40b91a21e293024f14404d5388efaa244cbd41cf58892f488cbdebf1a7d2"} Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.971336 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" event={"ID":"4ef05574-4924-4a73-b0ba-0e632c96410a","Type":"ContainerDied","Data":"4ad23a827a231477fdb77ba6ee3f420fc6df78c85d1f747cf3c87cf7c9d7423e"} Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.971388 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sqbkd" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.976741 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n6875"] Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.984213 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n6875"] Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.984859 4578 scope.go:117] "RemoveContainer" containerID="8f2dcd8ac4499b35a9c3fbdb82d3d415bb581b13a49e033da40e7f26f1567cc0" Oct 03 12:55:52 crc kubenswrapper[4578]: I1003 12:55:52.993717 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-bf7p9" podStartSLOduration=1.99369719 podStartE2EDuration="1.99369719s" podCreationTimestamp="2025-10-03 12:55:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:55:52.989066336 +0000 UTC m=+288.787538520" watchObservedRunningTime="2025-10-03 12:55:52.99369719 +0000 UTC m=+288.792169374" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.007276 4578 scope.go:117] "RemoveContainer" containerID="0eb7978f73b5b71a41f58d92e0989daa504e621cc70de06a65d413e9353b45bb" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.013852 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ldfq4"] Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.017440 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ldfq4"] Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.027445 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sqbkd"] Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.028745 4578 scope.go:117] "RemoveContainer" containerID="2e071a1b027ff081eb45821e9388f826ec1b24a2089965350e316344a307e6d9" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.035857 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sqbkd"] Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.045938 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5hjv"] Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.050486 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5hjv"] Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.055828 4578 scope.go:117] "RemoveContainer" containerID="5d83e6a2765fabe8666fefde1a7eca38a5cbad1ebfb98e191920c7b5afa511e0" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.074170 4578 scope.go:117] "RemoveContainer" containerID="c0443a230b0f1a59a152fd9fe5e5afab476c4c195bdbe35c0a33446f5e491377" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.093852 4578 scope.go:117] "RemoveContainer" containerID="5ef68c2cbc6ffb046975d4b63e43959fee1d0d2899a5b0a8878d2c1f40c51f76" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.107231 4578 scope.go:117] "RemoveContainer" containerID="bbf91797952d16a949c6632514a820d7c4fd06aeb5108932c7905c80d8358b78" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.120401 4578 scope.go:117] "RemoveContainer" containerID="439daae67e164ac422ee4754f63691531024d81e693b41b096aecd99ecedc43d" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.599203 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xtdbx"] Oct 03 12:55:53 crc kubenswrapper[4578]: E1003 12:55:53.599779 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75b24dc1-77ec-4b54-9f85-2443e273eadf" containerName="registry-server" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.599794 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="75b24dc1-77ec-4b54-9f85-2443e273eadf" containerName="registry-server" Oct 03 12:55:53 crc kubenswrapper[4578]: E1003 12:55:53.599808 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75b24dc1-77ec-4b54-9f85-2443e273eadf" containerName="extract-utilities" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.599816 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="75b24dc1-77ec-4b54-9f85-2443e273eadf" containerName="extract-utilities" Oct 03 12:55:53 crc kubenswrapper[4578]: E1003 12:55:53.599827 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75b24dc1-77ec-4b54-9f85-2443e273eadf" containerName="extract-content" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.599834 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="75b24dc1-77ec-4b54-9f85-2443e273eadf" containerName="extract-content" Oct 03 12:55:53 crc kubenswrapper[4578]: E1003 12:55:53.599845 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40b7869-c4f0-427d-b99c-fca7f0d8b937" containerName="extract-utilities" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.599852 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40b7869-c4f0-427d-b99c-fca7f0d8b937" containerName="extract-utilities" Oct 03 12:55:53 crc kubenswrapper[4578]: E1003 12:55:53.599861 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40b7869-c4f0-427d-b99c-fca7f0d8b937" containerName="extract-content" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.599868 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40b7869-c4f0-427d-b99c-fca7f0d8b937" containerName="extract-content" Oct 03 12:55:53 crc kubenswrapper[4578]: E1003 12:55:53.599881 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="742a4571-0a86-4cb4-bb93-8b46219f6d49" containerName="extract-content" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.599888 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="742a4571-0a86-4cb4-bb93-8b46219f6d49" containerName="extract-content" Oct 03 12:55:53 crc kubenswrapper[4578]: E1003 12:55:53.599896 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="742a4571-0a86-4cb4-bb93-8b46219f6d49" containerName="registry-server" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.599903 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="742a4571-0a86-4cb4-bb93-8b46219f6d49" containerName="registry-server" Oct 03 12:55:53 crc kubenswrapper[4578]: E1003 12:55:53.599912 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="742a4571-0a86-4cb4-bb93-8b46219f6d49" containerName="extract-utilities" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.599919 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="742a4571-0a86-4cb4-bb93-8b46219f6d49" containerName="extract-utilities" Oct 03 12:55:53 crc kubenswrapper[4578]: E1003 12:55:53.599929 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40b7869-c4f0-427d-b99c-fca7f0d8b937" containerName="registry-server" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.599936 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40b7869-c4f0-427d-b99c-fca7f0d8b937" containerName="registry-server" Oct 03 12:55:53 crc kubenswrapper[4578]: E1003 12:55:53.599961 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ef05574-4924-4a73-b0ba-0e632c96410a" containerName="marketplace-operator" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.599968 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ef05574-4924-4a73-b0ba-0e632c96410a" containerName="marketplace-operator" Oct 03 12:55:53 crc kubenswrapper[4578]: E1003 12:55:53.599980 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beb5b22e-cc9e-453b-bac2-88f7a5d30c29" containerName="registry-server" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.599987 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="beb5b22e-cc9e-453b-bac2-88f7a5d30c29" containerName="registry-server" Oct 03 12:55:53 crc kubenswrapper[4578]: E1003 12:55:53.599997 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beb5b22e-cc9e-453b-bac2-88f7a5d30c29" containerName="extract-content" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.600003 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="beb5b22e-cc9e-453b-bac2-88f7a5d30c29" containerName="extract-content" Oct 03 12:55:53 crc kubenswrapper[4578]: E1003 12:55:53.600014 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beb5b22e-cc9e-453b-bac2-88f7a5d30c29" containerName="extract-utilities" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.600021 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="beb5b22e-cc9e-453b-bac2-88f7a5d30c29" containerName="extract-utilities" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.600118 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="75b24dc1-77ec-4b54-9f85-2443e273eadf" containerName="registry-server" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.600131 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="742a4571-0a86-4cb4-bb93-8b46219f6d49" containerName="registry-server" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.600141 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="beb5b22e-cc9e-453b-bac2-88f7a5d30c29" containerName="registry-server" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.600151 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40b7869-c4f0-427d-b99c-fca7f0d8b937" containerName="registry-server" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.600159 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ef05574-4924-4a73-b0ba-0e632c96410a" containerName="marketplace-operator" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.601148 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xtdbx" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.603020 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.607618 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xtdbx"] Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.631306 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9-catalog-content\") pod \"redhat-marketplace-xtdbx\" (UID: \"90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9\") " pod="openshift-marketplace/redhat-marketplace-xtdbx" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.631395 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p96r8\" (UniqueName: \"kubernetes.io/projected/90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9-kube-api-access-p96r8\") pod \"redhat-marketplace-xtdbx\" (UID: \"90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9\") " pod="openshift-marketplace/redhat-marketplace-xtdbx" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.631417 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9-utilities\") pod \"redhat-marketplace-xtdbx\" (UID: \"90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9\") " pod="openshift-marketplace/redhat-marketplace-xtdbx" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.732320 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p96r8\" (UniqueName: \"kubernetes.io/projected/90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9-kube-api-access-p96r8\") pod \"redhat-marketplace-xtdbx\" (UID: \"90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9\") " pod="openshift-marketplace/redhat-marketplace-xtdbx" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.732370 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9-utilities\") pod \"redhat-marketplace-xtdbx\" (UID: \"90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9\") " pod="openshift-marketplace/redhat-marketplace-xtdbx" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.732442 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9-catalog-content\") pod \"redhat-marketplace-xtdbx\" (UID: \"90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9\") " pod="openshift-marketplace/redhat-marketplace-xtdbx" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.733046 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9-utilities\") pod \"redhat-marketplace-xtdbx\" (UID: \"90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9\") " pod="openshift-marketplace/redhat-marketplace-xtdbx" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.733069 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9-catalog-content\") pod \"redhat-marketplace-xtdbx\" (UID: \"90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9\") " pod="openshift-marketplace/redhat-marketplace-xtdbx" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.749969 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p96r8\" (UniqueName: \"kubernetes.io/projected/90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9-kube-api-access-p96r8\") pod \"redhat-marketplace-xtdbx\" (UID: \"90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9\") " pod="openshift-marketplace/redhat-marketplace-xtdbx" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.799848 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6wkpp"] Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.801041 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6wkpp" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.803255 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.814037 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6wkpp"] Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.833403 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68798069-f345-4517-88f7-d8d50c787b5d-catalog-content\") pod \"redhat-operators-6wkpp\" (UID: \"68798069-f345-4517-88f7-d8d50c787b5d\") " pod="openshift-marketplace/redhat-operators-6wkpp" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.833532 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68798069-f345-4517-88f7-d8d50c787b5d-utilities\") pod \"redhat-operators-6wkpp\" (UID: \"68798069-f345-4517-88f7-d8d50c787b5d\") " pod="openshift-marketplace/redhat-operators-6wkpp" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.833568 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rskt7\" (UniqueName: \"kubernetes.io/projected/68798069-f345-4517-88f7-d8d50c787b5d-kube-api-access-rskt7\") pod \"redhat-operators-6wkpp\" (UID: \"68798069-f345-4517-88f7-d8d50c787b5d\") " pod="openshift-marketplace/redhat-operators-6wkpp" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.928662 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xtdbx" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.935307 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68798069-f345-4517-88f7-d8d50c787b5d-utilities\") pod \"redhat-operators-6wkpp\" (UID: \"68798069-f345-4517-88f7-d8d50c787b5d\") " pod="openshift-marketplace/redhat-operators-6wkpp" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.935366 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rskt7\" (UniqueName: \"kubernetes.io/projected/68798069-f345-4517-88f7-d8d50c787b5d-kube-api-access-rskt7\") pod \"redhat-operators-6wkpp\" (UID: \"68798069-f345-4517-88f7-d8d50c787b5d\") " pod="openshift-marketplace/redhat-operators-6wkpp" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.935405 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68798069-f345-4517-88f7-d8d50c787b5d-catalog-content\") pod \"redhat-operators-6wkpp\" (UID: \"68798069-f345-4517-88f7-d8d50c787b5d\") " pod="openshift-marketplace/redhat-operators-6wkpp" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.935801 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68798069-f345-4517-88f7-d8d50c787b5d-catalog-content\") pod \"redhat-operators-6wkpp\" (UID: \"68798069-f345-4517-88f7-d8d50c787b5d\") " pod="openshift-marketplace/redhat-operators-6wkpp" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.935825 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68798069-f345-4517-88f7-d8d50c787b5d-utilities\") pod \"redhat-operators-6wkpp\" (UID: \"68798069-f345-4517-88f7-d8d50c787b5d\") " pod="openshift-marketplace/redhat-operators-6wkpp" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.956390 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rskt7\" (UniqueName: \"kubernetes.io/projected/68798069-f345-4517-88f7-d8d50c787b5d-kube-api-access-rskt7\") pod \"redhat-operators-6wkpp\" (UID: \"68798069-f345-4517-88f7-d8d50c787b5d\") " pod="openshift-marketplace/redhat-operators-6wkpp" Oct 03 12:55:53 crc kubenswrapper[4578]: I1003 12:55:53.988944 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-bf7p9" Oct 03 12:55:54 crc kubenswrapper[4578]: I1003 12:55:54.126948 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6wkpp" Oct 03 12:55:54 crc kubenswrapper[4578]: I1003 12:55:54.342935 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xtdbx"] Oct 03 12:55:54 crc kubenswrapper[4578]: I1003 12:55:54.533300 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6wkpp"] Oct 03 12:55:54 crc kubenswrapper[4578]: W1003 12:55:54.541709 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68798069_f345_4517_88f7_d8d50c787b5d.slice/crio-d4ebef1582cf21ac60331fa4e793801ff7956dea851ee1d791aed246112ab2e0 WatchSource:0}: Error finding container d4ebef1582cf21ac60331fa4e793801ff7956dea851ee1d791aed246112ab2e0: Status 404 returned error can't find the container with id d4ebef1582cf21ac60331fa4e793801ff7956dea851ee1d791aed246112ab2e0 Oct 03 12:55:54 crc kubenswrapper[4578]: I1003 12:55:54.914868 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ef05574-4924-4a73-b0ba-0e632c96410a" path="/var/lib/kubelet/pods/4ef05574-4924-4a73-b0ba-0e632c96410a/volumes" Oct 03 12:55:54 crc kubenswrapper[4578]: I1003 12:55:54.915750 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75b24dc1-77ec-4b54-9f85-2443e273eadf" path="/var/lib/kubelet/pods/75b24dc1-77ec-4b54-9f85-2443e273eadf/volumes" Oct 03 12:55:54 crc kubenswrapper[4578]: I1003 12:55:54.916432 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b40b7869-c4f0-427d-b99c-fca7f0d8b937" path="/var/lib/kubelet/pods/b40b7869-c4f0-427d-b99c-fca7f0d8b937/volumes" Oct 03 12:55:54 crc kubenswrapper[4578]: I1003 12:55:54.917820 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="beb5b22e-cc9e-453b-bac2-88f7a5d30c29" path="/var/lib/kubelet/pods/beb5b22e-cc9e-453b-bac2-88f7a5d30c29/volumes" Oct 03 12:55:54 crc kubenswrapper[4578]: I1003 12:55:54.991149 4578 generic.go:334] "Generic (PLEG): container finished" podID="90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9" containerID="32268016626dbf07a83d6d98fb73669db47442f22b961148e027e37c40315496" exitCode=0 Oct 03 12:55:54 crc kubenswrapper[4578]: I1003 12:55:54.991217 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xtdbx" event={"ID":"90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9","Type":"ContainerDied","Data":"32268016626dbf07a83d6d98fb73669db47442f22b961148e027e37c40315496"} Oct 03 12:55:54 crc kubenswrapper[4578]: I1003 12:55:54.991249 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xtdbx" event={"ID":"90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9","Type":"ContainerStarted","Data":"281f0434c685a741a814e802015248418271a6d0867b6d9cb47269e12269a07f"} Oct 03 12:55:54 crc kubenswrapper[4578]: I1003 12:55:54.994676 4578 generic.go:334] "Generic (PLEG): container finished" podID="68798069-f345-4517-88f7-d8d50c787b5d" containerID="1a926a599dea1ea44a459875b31160fec5ba13f91f1ced7a51bb03aaee4c43c0" exitCode=0 Oct 03 12:55:54 crc kubenswrapper[4578]: I1003 12:55:54.994797 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6wkpp" event={"ID":"68798069-f345-4517-88f7-d8d50c787b5d","Type":"ContainerDied","Data":"1a926a599dea1ea44a459875b31160fec5ba13f91f1ced7a51bb03aaee4c43c0"} Oct 03 12:55:54 crc kubenswrapper[4578]: I1003 12:55:54.994851 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6wkpp" event={"ID":"68798069-f345-4517-88f7-d8d50c787b5d","Type":"ContainerStarted","Data":"d4ebef1582cf21ac60331fa4e793801ff7956dea851ee1d791aed246112ab2e0"} Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.008698 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qz6ml"] Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.009900 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qz6ml" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.012353 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.017139 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qz6ml"] Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.066895 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqzml\" (UniqueName: \"kubernetes.io/projected/b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39-kube-api-access-tqzml\") pod \"community-operators-qz6ml\" (UID: \"b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39\") " pod="openshift-marketplace/community-operators-qz6ml" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.067403 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39-utilities\") pod \"community-operators-qz6ml\" (UID: \"b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39\") " pod="openshift-marketplace/community-operators-qz6ml" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.067537 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39-catalog-content\") pod \"community-operators-qz6ml\" (UID: \"b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39\") " pod="openshift-marketplace/community-operators-qz6ml" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.168464 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39-catalog-content\") pod \"community-operators-qz6ml\" (UID: \"b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39\") " pod="openshift-marketplace/community-operators-qz6ml" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.168782 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqzml\" (UniqueName: \"kubernetes.io/projected/b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39-kube-api-access-tqzml\") pod \"community-operators-qz6ml\" (UID: \"b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39\") " pod="openshift-marketplace/community-operators-qz6ml" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.169178 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39-utilities\") pod \"community-operators-qz6ml\" (UID: \"b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39\") " pod="openshift-marketplace/community-operators-qz6ml" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.169554 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39-utilities\") pod \"community-operators-qz6ml\" (UID: \"b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39\") " pod="openshift-marketplace/community-operators-qz6ml" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.169984 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39-catalog-content\") pod \"community-operators-qz6ml\" (UID: \"b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39\") " pod="openshift-marketplace/community-operators-qz6ml" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.189498 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqzml\" (UniqueName: \"kubernetes.io/projected/b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39-kube-api-access-tqzml\") pod \"community-operators-qz6ml\" (UID: \"b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39\") " pod="openshift-marketplace/community-operators-qz6ml" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.199791 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-x2b49"] Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.200996 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x2b49" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.204987 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.218600 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x2b49"] Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.270700 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5afc585-29cf-4658-8b97-a23d3e57f349-catalog-content\") pod \"certified-operators-x2b49\" (UID: \"d5afc585-29cf-4658-8b97-a23d3e57f349\") " pod="openshift-marketplace/certified-operators-x2b49" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.270754 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62p64\" (UniqueName: \"kubernetes.io/projected/d5afc585-29cf-4658-8b97-a23d3e57f349-kube-api-access-62p64\") pod \"certified-operators-x2b49\" (UID: \"d5afc585-29cf-4658-8b97-a23d3e57f349\") " pod="openshift-marketplace/certified-operators-x2b49" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.270820 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5afc585-29cf-4658-8b97-a23d3e57f349-utilities\") pod \"certified-operators-x2b49\" (UID: \"d5afc585-29cf-4658-8b97-a23d3e57f349\") " pod="openshift-marketplace/certified-operators-x2b49" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.327609 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qz6ml" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.372199 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5afc585-29cf-4658-8b97-a23d3e57f349-catalog-content\") pod \"certified-operators-x2b49\" (UID: \"d5afc585-29cf-4658-8b97-a23d3e57f349\") " pod="openshift-marketplace/certified-operators-x2b49" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.372525 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62p64\" (UniqueName: \"kubernetes.io/projected/d5afc585-29cf-4658-8b97-a23d3e57f349-kube-api-access-62p64\") pod \"certified-operators-x2b49\" (UID: \"d5afc585-29cf-4658-8b97-a23d3e57f349\") " pod="openshift-marketplace/certified-operators-x2b49" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.372714 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5afc585-29cf-4658-8b97-a23d3e57f349-utilities\") pod \"certified-operators-x2b49\" (UID: \"d5afc585-29cf-4658-8b97-a23d3e57f349\") " pod="openshift-marketplace/certified-operators-x2b49" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.373215 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5afc585-29cf-4658-8b97-a23d3e57f349-utilities\") pod \"certified-operators-x2b49\" (UID: \"d5afc585-29cf-4658-8b97-a23d3e57f349\") " pod="openshift-marketplace/certified-operators-x2b49" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.373229 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5afc585-29cf-4658-8b97-a23d3e57f349-catalog-content\") pod \"certified-operators-x2b49\" (UID: \"d5afc585-29cf-4658-8b97-a23d3e57f349\") " pod="openshift-marketplace/certified-operators-x2b49" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.393360 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62p64\" (UniqueName: \"kubernetes.io/projected/d5afc585-29cf-4658-8b97-a23d3e57f349-kube-api-access-62p64\") pod \"certified-operators-x2b49\" (UID: \"d5afc585-29cf-4658-8b97-a23d3e57f349\") " pod="openshift-marketplace/certified-operators-x2b49" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.525854 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x2b49" Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.734842 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qz6ml"] Oct 03 12:55:56 crc kubenswrapper[4578]: W1003 12:55:56.768257 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4c28ddc_2d83_4a6a_9a3b_64ebc9949d39.slice/crio-cc8885ab1f061c468b5e3ef5a4d45f26606ad6dca7db7835c58fd67bb8074fd9 WatchSource:0}: Error finding container cc8885ab1f061c468b5e3ef5a4d45f26606ad6dca7db7835c58fd67bb8074fd9: Status 404 returned error can't find the container with id cc8885ab1f061c468b5e3ef5a4d45f26606ad6dca7db7835c58fd67bb8074fd9 Oct 03 12:55:56 crc kubenswrapper[4578]: I1003 12:55:56.919106 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x2b49"] Oct 03 12:55:56 crc kubenswrapper[4578]: W1003 12:55:56.921217 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5afc585_29cf_4658_8b97_a23d3e57f349.slice/crio-34e0aff2d27ec0a2d57fe9f346465a58afd08541ca618d46e4316907f78ae660 WatchSource:0}: Error finding container 34e0aff2d27ec0a2d57fe9f346465a58afd08541ca618d46e4316907f78ae660: Status 404 returned error can't find the container with id 34e0aff2d27ec0a2d57fe9f346465a58afd08541ca618d46e4316907f78ae660 Oct 03 12:55:57 crc kubenswrapper[4578]: I1003 12:55:57.008655 4578 generic.go:334] "Generic (PLEG): container finished" podID="68798069-f345-4517-88f7-d8d50c787b5d" containerID="eb7c1de764737aa40c9593d95fd9bee1650f10f372e2c7ccdfb3a4fea3b8251a" exitCode=0 Oct 03 12:55:57 crc kubenswrapper[4578]: I1003 12:55:57.008731 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6wkpp" event={"ID":"68798069-f345-4517-88f7-d8d50c787b5d","Type":"ContainerDied","Data":"eb7c1de764737aa40c9593d95fd9bee1650f10f372e2c7ccdfb3a4fea3b8251a"} Oct 03 12:55:57 crc kubenswrapper[4578]: I1003 12:55:57.013228 4578 generic.go:334] "Generic (PLEG): container finished" podID="b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39" containerID="8511def5276b3dc60d6fdc319f88b14906fa80a802386ac6054cc71e1af9c3c2" exitCode=0 Oct 03 12:55:57 crc kubenswrapper[4578]: I1003 12:55:57.013291 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qz6ml" event={"ID":"b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39","Type":"ContainerDied","Data":"8511def5276b3dc60d6fdc319f88b14906fa80a802386ac6054cc71e1af9c3c2"} Oct 03 12:55:57 crc kubenswrapper[4578]: I1003 12:55:57.013316 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qz6ml" event={"ID":"b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39","Type":"ContainerStarted","Data":"cc8885ab1f061c468b5e3ef5a4d45f26606ad6dca7db7835c58fd67bb8074fd9"} Oct 03 12:55:57 crc kubenswrapper[4578]: I1003 12:55:57.017243 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2b49" event={"ID":"d5afc585-29cf-4658-8b97-a23d3e57f349","Type":"ContainerStarted","Data":"34e0aff2d27ec0a2d57fe9f346465a58afd08541ca618d46e4316907f78ae660"} Oct 03 12:55:57 crc kubenswrapper[4578]: I1003 12:55:57.022788 4578 generic.go:334] "Generic (PLEG): container finished" podID="90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9" containerID="fb6265a67b27fb789616c3f300294c56537d427c78daaf28730db9f237a4c2a4" exitCode=0 Oct 03 12:55:57 crc kubenswrapper[4578]: I1003 12:55:57.022901 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xtdbx" event={"ID":"90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9","Type":"ContainerDied","Data":"fb6265a67b27fb789616c3f300294c56537d427c78daaf28730db9f237a4c2a4"} Oct 03 12:55:58 crc kubenswrapper[4578]: I1003 12:55:58.029828 4578 generic.go:334] "Generic (PLEG): container finished" podID="d5afc585-29cf-4658-8b97-a23d3e57f349" containerID="7f9085a5a6652bc39bc85d9e76e78f013cbe4faf491c45f165d5b4ee754787a3" exitCode=0 Oct 03 12:55:58 crc kubenswrapper[4578]: I1003 12:55:58.029949 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2b49" event={"ID":"d5afc585-29cf-4658-8b97-a23d3e57f349","Type":"ContainerDied","Data":"7f9085a5a6652bc39bc85d9e76e78f013cbe4faf491c45f165d5b4ee754787a3"} Oct 03 12:55:58 crc kubenswrapper[4578]: I1003 12:55:58.035704 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6wkpp" event={"ID":"68798069-f345-4517-88f7-d8d50c787b5d","Type":"ContainerStarted","Data":"1fbbca7891564bb1b6dda4a4e82368514a19a54fe25fc3823764f6f3edf33e3f"} Oct 03 12:55:58 crc kubenswrapper[4578]: I1003 12:55:58.072812 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6wkpp" podStartSLOduration=2.377615664 podStartE2EDuration="5.072789137s" podCreationTimestamp="2025-10-03 12:55:53 +0000 UTC" firstStartedPulling="2025-10-03 12:55:54.996477202 +0000 UTC m=+290.794949386" lastFinishedPulling="2025-10-03 12:55:57.691650675 +0000 UTC m=+293.490122859" observedRunningTime="2025-10-03 12:55:58.069792978 +0000 UTC m=+293.868265172" watchObservedRunningTime="2025-10-03 12:55:58.072789137 +0000 UTC m=+293.871261321" Oct 03 12:55:59 crc kubenswrapper[4578]: I1003 12:55:59.054859 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xtdbx" event={"ID":"90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9","Type":"ContainerStarted","Data":"e12978962a3e1dc32d378508e82fea877dce1b850ace57fc6dc517a8f0688468"} Oct 03 12:55:59 crc kubenswrapper[4578]: I1003 12:55:59.056481 4578 generic.go:334] "Generic (PLEG): container finished" podID="b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39" containerID="0cbc26d61a2dcfdb9dadb4ab63f27cdd263c82d27b32c255a75f63de091e86d2" exitCode=0 Oct 03 12:55:59 crc kubenswrapper[4578]: I1003 12:55:59.056505 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qz6ml" event={"ID":"b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39","Type":"ContainerDied","Data":"0cbc26d61a2dcfdb9dadb4ab63f27cdd263c82d27b32c255a75f63de091e86d2"} Oct 03 12:55:59 crc kubenswrapper[4578]: I1003 12:55:59.058219 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2b49" event={"ID":"d5afc585-29cf-4658-8b97-a23d3e57f349","Type":"ContainerStarted","Data":"f518b9344b33e05af57e17333d789e978d57f9cc6410d35391a24ec76351ab7f"} Oct 03 12:55:59 crc kubenswrapper[4578]: I1003 12:55:59.072172 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xtdbx" podStartSLOduration=3.091008824 podStartE2EDuration="6.072156416s" podCreationTimestamp="2025-10-03 12:55:53 +0000 UTC" firstStartedPulling="2025-10-03 12:55:54.992389606 +0000 UTC m=+290.790861790" lastFinishedPulling="2025-10-03 12:55:57.973537198 +0000 UTC m=+293.772009382" observedRunningTime="2025-10-03 12:55:59.070734839 +0000 UTC m=+294.869207043" watchObservedRunningTime="2025-10-03 12:55:59.072156416 +0000 UTC m=+294.870628600" Oct 03 12:56:00 crc kubenswrapper[4578]: I1003 12:56:00.064219 4578 generic.go:334] "Generic (PLEG): container finished" podID="d5afc585-29cf-4658-8b97-a23d3e57f349" containerID="f518b9344b33e05af57e17333d789e978d57f9cc6410d35391a24ec76351ab7f" exitCode=0 Oct 03 12:56:00 crc kubenswrapper[4578]: I1003 12:56:00.064269 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2b49" event={"ID":"d5afc585-29cf-4658-8b97-a23d3e57f349","Type":"ContainerDied","Data":"f518b9344b33e05af57e17333d789e978d57f9cc6410d35391a24ec76351ab7f"} Oct 03 12:56:01 crc kubenswrapper[4578]: I1003 12:56:01.071015 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qz6ml" event={"ID":"b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39","Type":"ContainerStarted","Data":"8cfdba48eadd56ac0c75a47e5fb0a2c787a1fbffee57a1d3d10b775140fc436f"} Oct 03 12:56:01 crc kubenswrapper[4578]: I1003 12:56:01.073340 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2b49" event={"ID":"d5afc585-29cf-4658-8b97-a23d3e57f349","Type":"ContainerStarted","Data":"b2bb98a1b0ef323644a879966667b197d3247567d8f8710e9855a89b5282b901"} Oct 03 12:56:01 crc kubenswrapper[4578]: I1003 12:56:01.092008 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qz6ml" podStartSLOduration=3.658540935 podStartE2EDuration="6.091989856s" podCreationTimestamp="2025-10-03 12:55:55 +0000 UTC" firstStartedPulling="2025-10-03 12:55:57.018818154 +0000 UTC m=+292.817290338" lastFinishedPulling="2025-10-03 12:55:59.452267075 +0000 UTC m=+295.250739259" observedRunningTime="2025-10-03 12:56:01.09031997 +0000 UTC m=+296.888792164" watchObservedRunningTime="2025-10-03 12:56:01.091989856 +0000 UTC m=+296.890462040" Oct 03 12:56:03 crc kubenswrapper[4578]: I1003 12:56:03.929287 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xtdbx" Oct 03 12:56:03 crc kubenswrapper[4578]: I1003 12:56:03.929726 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xtdbx" Oct 03 12:56:03 crc kubenswrapper[4578]: I1003 12:56:03.974351 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xtdbx" Oct 03 12:56:03 crc kubenswrapper[4578]: I1003 12:56:03.993400 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-x2b49" podStartSLOduration=5.579468664 podStartE2EDuration="7.993381756s" podCreationTimestamp="2025-10-03 12:55:56 +0000 UTC" firstStartedPulling="2025-10-03 12:55:58.032500578 +0000 UTC m=+293.830972762" lastFinishedPulling="2025-10-03 12:56:00.44641367 +0000 UTC m=+296.244885854" observedRunningTime="2025-10-03 12:56:01.116661686 +0000 UTC m=+296.915133890" watchObservedRunningTime="2025-10-03 12:56:03.993381756 +0000 UTC m=+299.791853940" Oct 03 12:56:04 crc kubenswrapper[4578]: I1003 12:56:04.121738 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xtdbx" Oct 03 12:56:04 crc kubenswrapper[4578]: I1003 12:56:04.127667 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6wkpp" Oct 03 12:56:04 crc kubenswrapper[4578]: I1003 12:56:04.127708 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6wkpp" Oct 03 12:56:04 crc kubenswrapper[4578]: I1003 12:56:04.173189 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6wkpp" Oct 03 12:56:05 crc kubenswrapper[4578]: I1003 12:56:05.132554 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6wkpp" Oct 03 12:56:06 crc kubenswrapper[4578]: I1003 12:56:06.328937 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qz6ml" Oct 03 12:56:06 crc kubenswrapper[4578]: I1003 12:56:06.329332 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qz6ml" Oct 03 12:56:06 crc kubenswrapper[4578]: I1003 12:56:06.367541 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qz6ml" Oct 03 12:56:06 crc kubenswrapper[4578]: I1003 12:56:06.527379 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-x2b49" Oct 03 12:56:06 crc kubenswrapper[4578]: I1003 12:56:06.527737 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-x2b49" Oct 03 12:56:06 crc kubenswrapper[4578]: I1003 12:56:06.562167 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-x2b49" Oct 03 12:56:07 crc kubenswrapper[4578]: I1003 12:56:07.134694 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qz6ml" Oct 03 12:56:07 crc kubenswrapper[4578]: I1003 12:56:07.137270 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-x2b49" Oct 03 12:56:35 crc kubenswrapper[4578]: I1003 12:56:35.091266 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 12:56:35 crc kubenswrapper[4578]: I1003 12:56:35.092733 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 12:57:05 crc kubenswrapper[4578]: I1003 12:57:05.091794 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 12:57:05 crc kubenswrapper[4578]: I1003 12:57:05.092280 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 12:57:35 crc kubenswrapper[4578]: I1003 12:57:35.091182 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 12:57:35 crc kubenswrapper[4578]: I1003 12:57:35.091797 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 12:57:35 crc kubenswrapper[4578]: I1003 12:57:35.091851 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 12:57:35 crc kubenswrapper[4578]: I1003 12:57:35.092428 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"199792b4b113133fd7bee4bb1715ec98bb415c3fcc2500bb9a60f6c23d31f8a5"} pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 12:57:35 crc kubenswrapper[4578]: I1003 12:57:35.092480 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" containerID="cri-o://199792b4b113133fd7bee4bb1715ec98bb415c3fcc2500bb9a60f6c23d31f8a5" gracePeriod=600 Oct 03 12:57:35 crc kubenswrapper[4578]: I1003 12:57:35.528109 4578 generic.go:334] "Generic (PLEG): container finished" podID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerID="199792b4b113133fd7bee4bb1715ec98bb415c3fcc2500bb9a60f6c23d31f8a5" exitCode=0 Oct 03 12:57:35 crc kubenswrapper[4578]: I1003 12:57:35.528161 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerDied","Data":"199792b4b113133fd7bee4bb1715ec98bb415c3fcc2500bb9a60f6c23d31f8a5"} Oct 03 12:57:35 crc kubenswrapper[4578]: I1003 12:57:35.528376 4578 scope.go:117] "RemoveContainer" containerID="54d70463ff31b8bf3ba1a0554e9643cf32317a5a81b6b99dec103040a3688716" Oct 03 12:57:36 crc kubenswrapper[4578]: I1003 12:57:36.534233 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"9f7b9dc0d3d42c153f3e120deeab5edbba10dbc28e328d2398b250f0804ee378"} Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.246518 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-zk9dh"] Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.247949 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.263497 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-zk9dh"] Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.364483 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b1943423-91ed-4e23-a2e8-82c5c33480a5-installation-pull-secrets\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.364530 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b1943423-91ed-4e23-a2e8-82c5c33480a5-trusted-ca\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.364558 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b1943423-91ed-4e23-a2e8-82c5c33480a5-ca-trust-extracted\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.364578 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b1943423-91ed-4e23-a2e8-82c5c33480a5-bound-sa-token\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.364603 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.364667 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b1943423-91ed-4e23-a2e8-82c5c33480a5-registry-certificates\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.364707 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlwzb\" (UniqueName: \"kubernetes.io/projected/b1943423-91ed-4e23-a2e8-82c5c33480a5-kube-api-access-vlwzb\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.364733 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b1943423-91ed-4e23-a2e8-82c5c33480a5-registry-tls\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.399931 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.465205 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b1943423-91ed-4e23-a2e8-82c5c33480a5-installation-pull-secrets\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.465452 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b1943423-91ed-4e23-a2e8-82c5c33480a5-trusted-ca\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.465567 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b1943423-91ed-4e23-a2e8-82c5c33480a5-ca-trust-extracted\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.465671 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b1943423-91ed-4e23-a2e8-82c5c33480a5-bound-sa-token\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.465780 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b1943423-91ed-4e23-a2e8-82c5c33480a5-registry-certificates\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.466705 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlwzb\" (UniqueName: \"kubernetes.io/projected/b1943423-91ed-4e23-a2e8-82c5c33480a5-kube-api-access-vlwzb\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.466731 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b1943423-91ed-4e23-a2e8-82c5c33480a5-registry-tls\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.466064 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b1943423-91ed-4e23-a2e8-82c5c33480a5-ca-trust-extracted\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.467117 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b1943423-91ed-4e23-a2e8-82c5c33480a5-registry-certificates\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.467373 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b1943423-91ed-4e23-a2e8-82c5c33480a5-trusted-ca\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.470724 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b1943423-91ed-4e23-a2e8-82c5c33480a5-installation-pull-secrets\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.470760 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b1943423-91ed-4e23-a2e8-82c5c33480a5-registry-tls\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.482968 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b1943423-91ed-4e23-a2e8-82c5c33480a5-bound-sa-token\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.487337 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlwzb\" (UniqueName: \"kubernetes.io/projected/b1943423-91ed-4e23-a2e8-82c5c33480a5-kube-api-access-vlwzb\") pod \"image-registry-66df7c8f76-zk9dh\" (UID: \"b1943423-91ed-4e23-a2e8-82c5c33480a5\") " pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.563022 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:30 crc kubenswrapper[4578]: I1003 12:59:30.729568 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-zk9dh"] Oct 03 12:59:31 crc kubenswrapper[4578]: I1003 12:59:31.102521 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" event={"ID":"b1943423-91ed-4e23-a2e8-82c5c33480a5","Type":"ContainerStarted","Data":"e7aa3af256a4e46acf2bfa54a18ead682ee295bc559488a3f9e0e98e6a294ee9"} Oct 03 12:59:31 crc kubenswrapper[4578]: I1003 12:59:31.102952 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" event={"ID":"b1943423-91ed-4e23-a2e8-82c5c33480a5","Type":"ContainerStarted","Data":"993267344fefb26e975e7d053b35d1e03b11be18433840690014755761540505"} Oct 03 12:59:32 crc kubenswrapper[4578]: I1003 12:59:32.108458 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:32 crc kubenswrapper[4578]: I1003 12:59:32.143813 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" podStartSLOduration=2.143788339 podStartE2EDuration="2.143788339s" podCreationTimestamp="2025-10-03 12:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 12:59:32.133833387 +0000 UTC m=+507.932305601" watchObservedRunningTime="2025-10-03 12:59:32.143788339 +0000 UTC m=+507.942260533" Oct 03 12:59:35 crc kubenswrapper[4578]: I1003 12:59:35.091543 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 12:59:35 crc kubenswrapper[4578]: I1003 12:59:35.092063 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 12:59:50 crc kubenswrapper[4578]: I1003 12:59:50.569272 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-zk9dh" Oct 03 12:59:50 crc kubenswrapper[4578]: I1003 12:59:50.630945 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5xxhk"] Oct 03 13:00:00 crc kubenswrapper[4578]: I1003 13:00:00.136041 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z"] Oct 03 13:00:00 crc kubenswrapper[4578]: I1003 13:00:00.138335 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z" Oct 03 13:00:00 crc kubenswrapper[4578]: I1003 13:00:00.144436 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 13:00:00 crc kubenswrapper[4578]: I1003 13:00:00.144822 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 13:00:00 crc kubenswrapper[4578]: I1003 13:00:00.147337 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z"] Oct 03 13:00:00 crc kubenswrapper[4578]: I1003 13:00:00.247006 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqvfl\" (UniqueName: \"kubernetes.io/projected/d777f4d5-fab6-4033-9200-e7cce416dab9-kube-api-access-lqvfl\") pod \"collect-profiles-29324940-gpv7z\" (UID: \"d777f4d5-fab6-4033-9200-e7cce416dab9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z" Oct 03 13:00:00 crc kubenswrapper[4578]: I1003 13:00:00.247069 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d777f4d5-fab6-4033-9200-e7cce416dab9-config-volume\") pod \"collect-profiles-29324940-gpv7z\" (UID: \"d777f4d5-fab6-4033-9200-e7cce416dab9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z" Oct 03 13:00:00 crc kubenswrapper[4578]: I1003 13:00:00.247139 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d777f4d5-fab6-4033-9200-e7cce416dab9-secret-volume\") pod \"collect-profiles-29324940-gpv7z\" (UID: \"d777f4d5-fab6-4033-9200-e7cce416dab9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z" Oct 03 13:00:00 crc kubenswrapper[4578]: I1003 13:00:00.348027 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqvfl\" (UniqueName: \"kubernetes.io/projected/d777f4d5-fab6-4033-9200-e7cce416dab9-kube-api-access-lqvfl\") pod \"collect-profiles-29324940-gpv7z\" (UID: \"d777f4d5-fab6-4033-9200-e7cce416dab9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z" Oct 03 13:00:00 crc kubenswrapper[4578]: I1003 13:00:00.348346 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d777f4d5-fab6-4033-9200-e7cce416dab9-config-volume\") pod \"collect-profiles-29324940-gpv7z\" (UID: \"d777f4d5-fab6-4033-9200-e7cce416dab9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z" Oct 03 13:00:00 crc kubenswrapper[4578]: I1003 13:00:00.348445 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d777f4d5-fab6-4033-9200-e7cce416dab9-secret-volume\") pod \"collect-profiles-29324940-gpv7z\" (UID: \"d777f4d5-fab6-4033-9200-e7cce416dab9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z" Oct 03 13:00:00 crc kubenswrapper[4578]: I1003 13:00:00.349308 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d777f4d5-fab6-4033-9200-e7cce416dab9-config-volume\") pod \"collect-profiles-29324940-gpv7z\" (UID: \"d777f4d5-fab6-4033-9200-e7cce416dab9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z" Oct 03 13:00:00 crc kubenswrapper[4578]: I1003 13:00:00.368354 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d777f4d5-fab6-4033-9200-e7cce416dab9-secret-volume\") pod \"collect-profiles-29324940-gpv7z\" (UID: \"d777f4d5-fab6-4033-9200-e7cce416dab9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z" Oct 03 13:00:00 crc kubenswrapper[4578]: I1003 13:00:00.371779 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqvfl\" (UniqueName: \"kubernetes.io/projected/d777f4d5-fab6-4033-9200-e7cce416dab9-kube-api-access-lqvfl\") pod \"collect-profiles-29324940-gpv7z\" (UID: \"d777f4d5-fab6-4033-9200-e7cce416dab9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z" Oct 03 13:00:00 crc kubenswrapper[4578]: I1003 13:00:00.464453 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z" Oct 03 13:00:00 crc kubenswrapper[4578]: I1003 13:00:00.635938 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z"] Oct 03 13:00:01 crc kubenswrapper[4578]: I1003 13:00:01.246687 4578 generic.go:334] "Generic (PLEG): container finished" podID="d777f4d5-fab6-4033-9200-e7cce416dab9" containerID="7587bf723aa48d53704b41dd167508e205437c8b6ed9ac61bf3b4b21fdd01e82" exitCode=0 Oct 03 13:00:01 crc kubenswrapper[4578]: I1003 13:00:01.246802 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z" event={"ID":"d777f4d5-fab6-4033-9200-e7cce416dab9","Type":"ContainerDied","Data":"7587bf723aa48d53704b41dd167508e205437c8b6ed9ac61bf3b4b21fdd01e82"} Oct 03 13:00:01 crc kubenswrapper[4578]: I1003 13:00:01.247520 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z" event={"ID":"d777f4d5-fab6-4033-9200-e7cce416dab9","Type":"ContainerStarted","Data":"50dcb44c4513f1998ff0f7cfe7ecc41fa39b6b149aed29f9e617d841ef521e4d"} Oct 03 13:00:02 crc kubenswrapper[4578]: I1003 13:00:02.488820 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z" Oct 03 13:00:02 crc kubenswrapper[4578]: I1003 13:00:02.577935 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d777f4d5-fab6-4033-9200-e7cce416dab9-secret-volume\") pod \"d777f4d5-fab6-4033-9200-e7cce416dab9\" (UID: \"d777f4d5-fab6-4033-9200-e7cce416dab9\") " Oct 03 13:00:02 crc kubenswrapper[4578]: I1003 13:00:02.578046 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqvfl\" (UniqueName: \"kubernetes.io/projected/d777f4d5-fab6-4033-9200-e7cce416dab9-kube-api-access-lqvfl\") pod \"d777f4d5-fab6-4033-9200-e7cce416dab9\" (UID: \"d777f4d5-fab6-4033-9200-e7cce416dab9\") " Oct 03 13:00:02 crc kubenswrapper[4578]: I1003 13:00:02.578096 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d777f4d5-fab6-4033-9200-e7cce416dab9-config-volume\") pod \"d777f4d5-fab6-4033-9200-e7cce416dab9\" (UID: \"d777f4d5-fab6-4033-9200-e7cce416dab9\") " Oct 03 13:00:02 crc kubenswrapper[4578]: I1003 13:00:02.578714 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d777f4d5-fab6-4033-9200-e7cce416dab9-config-volume" (OuterVolumeSpecName: "config-volume") pod "d777f4d5-fab6-4033-9200-e7cce416dab9" (UID: "d777f4d5-fab6-4033-9200-e7cce416dab9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:00:02 crc kubenswrapper[4578]: I1003 13:00:02.582611 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d777f4d5-fab6-4033-9200-e7cce416dab9-kube-api-access-lqvfl" (OuterVolumeSpecName: "kube-api-access-lqvfl") pod "d777f4d5-fab6-4033-9200-e7cce416dab9" (UID: "d777f4d5-fab6-4033-9200-e7cce416dab9"). InnerVolumeSpecName "kube-api-access-lqvfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:00:02 crc kubenswrapper[4578]: I1003 13:00:02.582863 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d777f4d5-fab6-4033-9200-e7cce416dab9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d777f4d5-fab6-4033-9200-e7cce416dab9" (UID: "d777f4d5-fab6-4033-9200-e7cce416dab9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:00:02 crc kubenswrapper[4578]: I1003 13:00:02.679034 4578 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d777f4d5-fab6-4033-9200-e7cce416dab9-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 13:00:02 crc kubenswrapper[4578]: I1003 13:00:02.679067 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqvfl\" (UniqueName: \"kubernetes.io/projected/d777f4d5-fab6-4033-9200-e7cce416dab9-kube-api-access-lqvfl\") on node \"crc\" DevicePath \"\"" Oct 03 13:00:02 crc kubenswrapper[4578]: I1003 13:00:02.679076 4578 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d777f4d5-fab6-4033-9200-e7cce416dab9-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 13:00:03 crc kubenswrapper[4578]: I1003 13:00:03.261984 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z" event={"ID":"d777f4d5-fab6-4033-9200-e7cce416dab9","Type":"ContainerDied","Data":"50dcb44c4513f1998ff0f7cfe7ecc41fa39b6b149aed29f9e617d841ef521e4d"} Oct 03 13:00:03 crc kubenswrapper[4578]: I1003 13:00:03.262026 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50dcb44c4513f1998ff0f7cfe7ecc41fa39b6b149aed29f9e617d841ef521e4d" Oct 03 13:00:03 crc kubenswrapper[4578]: I1003 13:00:03.262030 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z" Oct 03 13:00:05 crc kubenswrapper[4578]: I1003 13:00:05.091311 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:00:05 crc kubenswrapper[4578]: I1003 13:00:05.091772 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:00:15 crc kubenswrapper[4578]: I1003 13:00:15.672321 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" podUID="1a1b400e-f3c1-456c-9d7b-0639a86edbd5" containerName="registry" containerID="cri-o://2c4a1cd87572a4535793447207bc93408844637dfbc08983920f83393b06d301" gracePeriod=30 Oct 03 13:00:15 crc kubenswrapper[4578]: I1003 13:00:15.975512 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.047576 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-ca-trust-extracted\") pod \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.047701 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-registry-tls\") pod \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.047778 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-bound-sa-token\") pod \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.047874 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-registry-certificates\") pod \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.048138 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.048192 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8h85\" (UniqueName: \"kubernetes.io/projected/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-kube-api-access-f8h85\") pod \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.048241 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-installation-pull-secrets\") pod \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.048298 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-trusted-ca\") pod \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\" (UID: \"1a1b400e-f3c1-456c-9d7b-0639a86edbd5\") " Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.049566 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "1a1b400e-f3c1-456c-9d7b-0639a86edbd5" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.050267 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "1a1b400e-f3c1-456c-9d7b-0639a86edbd5" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.055878 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "1a1b400e-f3c1-456c-9d7b-0639a86edbd5" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.056081 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "1a1b400e-f3c1-456c-9d7b-0639a86edbd5" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.056364 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-kube-api-access-f8h85" (OuterVolumeSpecName: "kube-api-access-f8h85") pod "1a1b400e-f3c1-456c-9d7b-0639a86edbd5" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5"). InnerVolumeSpecName "kube-api-access-f8h85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.056700 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "1a1b400e-f3c1-456c-9d7b-0639a86edbd5" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.060956 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "1a1b400e-f3c1-456c-9d7b-0639a86edbd5" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.064464 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "1a1b400e-f3c1-456c-9d7b-0639a86edbd5" (UID: "1a1b400e-f3c1-456c-9d7b-0639a86edbd5"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.149677 4578 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.149712 4578 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.149723 4578 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.149733 4578 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.149742 4578 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.149751 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8h85\" (UniqueName: \"kubernetes.io/projected/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-kube-api-access-f8h85\") on node \"crc\" DevicePath \"\"" Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.149760 4578 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1a1b400e-f3c1-456c-9d7b-0639a86edbd5-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.318867 4578 generic.go:334] "Generic (PLEG): container finished" podID="1a1b400e-f3c1-456c-9d7b-0639a86edbd5" containerID="2c4a1cd87572a4535793447207bc93408844637dfbc08983920f83393b06d301" exitCode=0 Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.318913 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" event={"ID":"1a1b400e-f3c1-456c-9d7b-0639a86edbd5","Type":"ContainerDied","Data":"2c4a1cd87572a4535793447207bc93408844637dfbc08983920f83393b06d301"} Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.318945 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" event={"ID":"1a1b400e-f3c1-456c-9d7b-0639a86edbd5","Type":"ContainerDied","Data":"23665e827ecb8df70b8f79e865d560acbb2df4c104887ea3cbf3c3b338579f3d"} Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.318960 4578 scope.go:117] "RemoveContainer" containerID="2c4a1cd87572a4535793447207bc93408844637dfbc08983920f83393b06d301" Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.319158 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5xxhk" Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.337072 4578 scope.go:117] "RemoveContainer" containerID="2c4a1cd87572a4535793447207bc93408844637dfbc08983920f83393b06d301" Oct 03 13:00:16 crc kubenswrapper[4578]: E1003 13:00:16.337576 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c4a1cd87572a4535793447207bc93408844637dfbc08983920f83393b06d301\": container with ID starting with 2c4a1cd87572a4535793447207bc93408844637dfbc08983920f83393b06d301 not found: ID does not exist" containerID="2c4a1cd87572a4535793447207bc93408844637dfbc08983920f83393b06d301" Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.337607 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c4a1cd87572a4535793447207bc93408844637dfbc08983920f83393b06d301"} err="failed to get container status \"2c4a1cd87572a4535793447207bc93408844637dfbc08983920f83393b06d301\": rpc error: code = NotFound desc = could not find container \"2c4a1cd87572a4535793447207bc93408844637dfbc08983920f83393b06d301\": container with ID starting with 2c4a1cd87572a4535793447207bc93408844637dfbc08983920f83393b06d301 not found: ID does not exist" Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.355959 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5xxhk"] Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.362996 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5xxhk"] Oct 03 13:00:16 crc kubenswrapper[4578]: I1003 13:00:16.915223 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a1b400e-f3c1-456c-9d7b-0639a86edbd5" path="/var/lib/kubelet/pods/1a1b400e-f3c1-456c-9d7b-0639a86edbd5/volumes" Oct 03 13:00:35 crc kubenswrapper[4578]: I1003 13:00:35.091243 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:00:35 crc kubenswrapper[4578]: I1003 13:00:35.091673 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:00:35 crc kubenswrapper[4578]: I1003 13:00:35.091717 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 13:00:35 crc kubenswrapper[4578]: I1003 13:00:35.092099 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9f7b9dc0d3d42c153f3e120deeab5edbba10dbc28e328d2398b250f0804ee378"} pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:00:35 crc kubenswrapper[4578]: I1003 13:00:35.092138 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" containerID="cri-o://9f7b9dc0d3d42c153f3e120deeab5edbba10dbc28e328d2398b250f0804ee378" gracePeriod=600 Oct 03 13:00:35 crc kubenswrapper[4578]: I1003 13:00:35.407627 4578 generic.go:334] "Generic (PLEG): container finished" podID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerID="9f7b9dc0d3d42c153f3e120deeab5edbba10dbc28e328d2398b250f0804ee378" exitCode=0 Oct 03 13:00:35 crc kubenswrapper[4578]: I1003 13:00:35.407676 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerDied","Data":"9f7b9dc0d3d42c153f3e120deeab5edbba10dbc28e328d2398b250f0804ee378"} Oct 03 13:00:35 crc kubenswrapper[4578]: I1003 13:00:35.407992 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"9fca9ac685f413e21ab5da95ad3fbaecf2db649af27947616da386bcc758120f"} Oct 03 13:00:35 crc kubenswrapper[4578]: I1003 13:00:35.408018 4578 scope.go:117] "RemoveContainer" containerID="199792b4b113133fd7bee4bb1715ec98bb415c3fcc2500bb9a60f6c23d31f8a5" Oct 03 13:02:35 crc kubenswrapper[4578]: I1003 13:02:35.091742 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:02:35 crc kubenswrapper[4578]: I1003 13:02:35.092385 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:03:05 crc kubenswrapper[4578]: I1003 13:03:05.091389 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:03:05 crc kubenswrapper[4578]: I1003 13:03:05.091901 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:03:35 crc kubenswrapper[4578]: I1003 13:03:35.091783 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:03:35 crc kubenswrapper[4578]: I1003 13:03:35.092327 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:03:35 crc kubenswrapper[4578]: I1003 13:03:35.092377 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 13:03:35 crc kubenswrapper[4578]: I1003 13:03:35.092964 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9fca9ac685f413e21ab5da95ad3fbaecf2db649af27947616da386bcc758120f"} pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:03:35 crc kubenswrapper[4578]: I1003 13:03:35.093018 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" containerID="cri-o://9fca9ac685f413e21ab5da95ad3fbaecf2db649af27947616da386bcc758120f" gracePeriod=600 Oct 03 13:03:35 crc kubenswrapper[4578]: I1003 13:03:35.265685 4578 generic.go:334] "Generic (PLEG): container finished" podID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerID="9fca9ac685f413e21ab5da95ad3fbaecf2db649af27947616da386bcc758120f" exitCode=0 Oct 03 13:03:35 crc kubenswrapper[4578]: I1003 13:03:35.265731 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerDied","Data":"9fca9ac685f413e21ab5da95ad3fbaecf2db649af27947616da386bcc758120f"} Oct 03 13:03:35 crc kubenswrapper[4578]: I1003 13:03:35.265768 4578 scope.go:117] "RemoveContainer" containerID="9f7b9dc0d3d42c153f3e120deeab5edbba10dbc28e328d2398b250f0804ee378" Oct 03 13:03:36 crc kubenswrapper[4578]: I1003 13:03:36.272542 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"942bca54ab772aabe22d68921c64f6de2fc3db7407fb9842d79fce2fa9613a96"} Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.316406 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hh7tq"] Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.318264 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" podUID="b2ee8da6-43e0-496d-aea1-2c7849305419" containerName="controller-manager" containerID="cri-o://543baf20b344ae188ff98b93c23363072c8aafa779ad8e8e546be787bb6c7df1" gracePeriod=30 Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.416902 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n"] Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.417463 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" podUID="5a1d0b1c-9bae-483f-886d-bf4acffee7d9" containerName="route-controller-manager" containerID="cri-o://b1dda7af6c7a57e9ca017010f0339fcf1e4b44c91b868d9c9f7093414c9538b5" gracePeriod=30 Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.683893 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.800487 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.859181 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2ee8da6-43e0-496d-aea1-2c7849305419-serving-cert\") pod \"b2ee8da6-43e0-496d-aea1-2c7849305419\" (UID: \"b2ee8da6-43e0-496d-aea1-2c7849305419\") " Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.859236 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g28ws\" (UniqueName: \"kubernetes.io/projected/b2ee8da6-43e0-496d-aea1-2c7849305419-kube-api-access-g28ws\") pod \"b2ee8da6-43e0-496d-aea1-2c7849305419\" (UID: \"b2ee8da6-43e0-496d-aea1-2c7849305419\") " Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.859286 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2ee8da6-43e0-496d-aea1-2c7849305419-config\") pod \"b2ee8da6-43e0-496d-aea1-2c7849305419\" (UID: \"b2ee8da6-43e0-496d-aea1-2c7849305419\") " Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.859314 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b2ee8da6-43e0-496d-aea1-2c7849305419-client-ca\") pod \"b2ee8da6-43e0-496d-aea1-2c7849305419\" (UID: \"b2ee8da6-43e0-496d-aea1-2c7849305419\") " Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.859368 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b2ee8da6-43e0-496d-aea1-2c7849305419-proxy-ca-bundles\") pod \"b2ee8da6-43e0-496d-aea1-2c7849305419\" (UID: \"b2ee8da6-43e0-496d-aea1-2c7849305419\") " Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.860204 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2ee8da6-43e0-496d-aea1-2c7849305419-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b2ee8da6-43e0-496d-aea1-2c7849305419" (UID: "b2ee8da6-43e0-496d-aea1-2c7849305419"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.860652 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2ee8da6-43e0-496d-aea1-2c7849305419-config" (OuterVolumeSpecName: "config") pod "b2ee8da6-43e0-496d-aea1-2c7849305419" (UID: "b2ee8da6-43e0-496d-aea1-2c7849305419"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.861056 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2ee8da6-43e0-496d-aea1-2c7849305419-client-ca" (OuterVolumeSpecName: "client-ca") pod "b2ee8da6-43e0-496d-aea1-2c7849305419" (UID: "b2ee8da6-43e0-496d-aea1-2c7849305419"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.866388 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2ee8da6-43e0-496d-aea1-2c7849305419-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b2ee8da6-43e0-496d-aea1-2c7849305419" (UID: "b2ee8da6-43e0-496d-aea1-2c7849305419"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.866495 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2ee8da6-43e0-496d-aea1-2c7849305419-kube-api-access-g28ws" (OuterVolumeSpecName: "kube-api-access-g28ws") pod "b2ee8da6-43e0-496d-aea1-2c7849305419" (UID: "b2ee8da6-43e0-496d-aea1-2c7849305419"). InnerVolumeSpecName "kube-api-access-g28ws". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.960105 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-serving-cert\") pod \"5a1d0b1c-9bae-483f-886d-bf4acffee7d9\" (UID: \"5a1d0b1c-9bae-483f-886d-bf4acffee7d9\") " Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.960173 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9mzs\" (UniqueName: \"kubernetes.io/projected/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-kube-api-access-h9mzs\") pod \"5a1d0b1c-9bae-483f-886d-bf4acffee7d9\" (UID: \"5a1d0b1c-9bae-483f-886d-bf4acffee7d9\") " Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.960240 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-config\") pod \"5a1d0b1c-9bae-483f-886d-bf4acffee7d9\" (UID: \"5a1d0b1c-9bae-483f-886d-bf4acffee7d9\") " Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.960257 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-client-ca\") pod \"5a1d0b1c-9bae-483f-886d-bf4acffee7d9\" (UID: \"5a1d0b1c-9bae-483f-886d-bf4acffee7d9\") " Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.960462 4578 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b2ee8da6-43e0-496d-aea1-2c7849305419-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.960482 4578 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2ee8da6-43e0-496d-aea1-2c7849305419-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.960495 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g28ws\" (UniqueName: \"kubernetes.io/projected/b2ee8da6-43e0-496d-aea1-2c7849305419-kube-api-access-g28ws\") on node \"crc\" DevicePath \"\"" Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.960509 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2ee8da6-43e0-496d-aea1-2c7849305419-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.960519 4578 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b2ee8da6-43e0-496d-aea1-2c7849305419-client-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.960982 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5a1d0b1c-9bae-483f-886d-bf4acffee7d9" (UID: "5a1d0b1c-9bae-483f-886d-bf4acffee7d9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.960997 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-config" (OuterVolumeSpecName: "config") pod "5a1d0b1c-9bae-483f-886d-bf4acffee7d9" (UID: "5a1d0b1c-9bae-483f-886d-bf4acffee7d9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.963322 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5a1d0b1c-9bae-483f-886d-bf4acffee7d9" (UID: "5a1d0b1c-9bae-483f-886d-bf4acffee7d9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:03:46 crc kubenswrapper[4578]: I1003 13:03:46.963399 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-kube-api-access-h9mzs" (OuterVolumeSpecName: "kube-api-access-h9mzs") pod "5a1d0b1c-9bae-483f-886d-bf4acffee7d9" (UID: "5a1d0b1c-9bae-483f-886d-bf4acffee7d9"). InnerVolumeSpecName "kube-api-access-h9mzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.061865 4578 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.061901 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9mzs\" (UniqueName: \"kubernetes.io/projected/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-kube-api-access-h9mzs\") on node \"crc\" DevicePath \"\"" Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.061916 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.061928 4578 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5a1d0b1c-9bae-483f-886d-bf4acffee7d9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.323519 4578 generic.go:334] "Generic (PLEG): container finished" podID="b2ee8da6-43e0-496d-aea1-2c7849305419" containerID="543baf20b344ae188ff98b93c23363072c8aafa779ad8e8e546be787bb6c7df1" exitCode=0 Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.323596 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.323660 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" event={"ID":"b2ee8da6-43e0-496d-aea1-2c7849305419","Type":"ContainerDied","Data":"543baf20b344ae188ff98b93c23363072c8aafa779ad8e8e546be787bb6c7df1"} Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.323735 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hh7tq" event={"ID":"b2ee8da6-43e0-496d-aea1-2c7849305419","Type":"ContainerDied","Data":"a0e3082a18351f620c4121d58b0d8d6d395a7706e4c6e90ebf3d8ced17b5fcb6"} Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.323753 4578 scope.go:117] "RemoveContainer" containerID="543baf20b344ae188ff98b93c23363072c8aafa779ad8e8e546be787bb6c7df1" Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.326652 4578 generic.go:334] "Generic (PLEG): container finished" podID="5a1d0b1c-9bae-483f-886d-bf4acffee7d9" containerID="b1dda7af6c7a57e9ca017010f0339fcf1e4b44c91b868d9c9f7093414c9538b5" exitCode=0 Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.326688 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" event={"ID":"5a1d0b1c-9bae-483f-886d-bf4acffee7d9","Type":"ContainerDied","Data":"b1dda7af6c7a57e9ca017010f0339fcf1e4b44c91b868d9c9f7093414c9538b5"} Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.326717 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" event={"ID":"5a1d0b1c-9bae-483f-886d-bf4acffee7d9","Type":"ContainerDied","Data":"e691b4315d1f3f216f70d30ef2ea86c360874b0609153fb8a175555365b37074"} Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.326852 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n" Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.345730 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hh7tq"] Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.345807 4578 scope.go:117] "RemoveContainer" containerID="543baf20b344ae188ff98b93c23363072c8aafa779ad8e8e546be787bb6c7df1" Oct 03 13:03:47 crc kubenswrapper[4578]: E1003 13:03:47.346505 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"543baf20b344ae188ff98b93c23363072c8aafa779ad8e8e546be787bb6c7df1\": container with ID starting with 543baf20b344ae188ff98b93c23363072c8aafa779ad8e8e546be787bb6c7df1 not found: ID does not exist" containerID="543baf20b344ae188ff98b93c23363072c8aafa779ad8e8e546be787bb6c7df1" Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.346580 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"543baf20b344ae188ff98b93c23363072c8aafa779ad8e8e546be787bb6c7df1"} err="failed to get container status \"543baf20b344ae188ff98b93c23363072c8aafa779ad8e8e546be787bb6c7df1\": rpc error: code = NotFound desc = could not find container \"543baf20b344ae188ff98b93c23363072c8aafa779ad8e8e546be787bb6c7df1\": container with ID starting with 543baf20b344ae188ff98b93c23363072c8aafa779ad8e8e546be787bb6c7df1 not found: ID does not exist" Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.346615 4578 scope.go:117] "RemoveContainer" containerID="b1dda7af6c7a57e9ca017010f0339fcf1e4b44c91b868d9c9f7093414c9538b5" Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.349841 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hh7tq"] Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.361569 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n"] Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.365723 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-t9q7n"] Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.369571 4578 scope.go:117] "RemoveContainer" containerID="b1dda7af6c7a57e9ca017010f0339fcf1e4b44c91b868d9c9f7093414c9538b5" Oct 03 13:03:47 crc kubenswrapper[4578]: E1003 13:03:47.370138 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1dda7af6c7a57e9ca017010f0339fcf1e4b44c91b868d9c9f7093414c9538b5\": container with ID starting with b1dda7af6c7a57e9ca017010f0339fcf1e4b44c91b868d9c9f7093414c9538b5 not found: ID does not exist" containerID="b1dda7af6c7a57e9ca017010f0339fcf1e4b44c91b868d9c9f7093414c9538b5" Oct 03 13:03:47 crc kubenswrapper[4578]: I1003 13:03:47.370170 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1dda7af6c7a57e9ca017010f0339fcf1e4b44c91b868d9c9f7093414c9538b5"} err="failed to get container status \"b1dda7af6c7a57e9ca017010f0339fcf1e4b44c91b868d9c9f7093414c9538b5\": rpc error: code = NotFound desc = could not find container \"b1dda7af6c7a57e9ca017010f0339fcf1e4b44c91b868d9c9f7093414c9538b5\": container with ID starting with b1dda7af6c7a57e9ca017010f0339fcf1e4b44c91b868d9c9f7093414c9538b5 not found: ID does not exist" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.134539 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p"] Oct 03 13:03:48 crc kubenswrapper[4578]: E1003 13:03:48.134743 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ee8da6-43e0-496d-aea1-2c7849305419" containerName="controller-manager" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.134755 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ee8da6-43e0-496d-aea1-2c7849305419" containerName="controller-manager" Oct 03 13:03:48 crc kubenswrapper[4578]: E1003 13:03:48.134771 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a1d0b1c-9bae-483f-886d-bf4acffee7d9" containerName="route-controller-manager" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.134777 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a1d0b1c-9bae-483f-886d-bf4acffee7d9" containerName="route-controller-manager" Oct 03 13:03:48 crc kubenswrapper[4578]: E1003 13:03:48.134788 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d777f4d5-fab6-4033-9200-e7cce416dab9" containerName="collect-profiles" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.134795 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="d777f4d5-fab6-4033-9200-e7cce416dab9" containerName="collect-profiles" Oct 03 13:03:48 crc kubenswrapper[4578]: E1003 13:03:48.134806 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a1b400e-f3c1-456c-9d7b-0639a86edbd5" containerName="registry" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.134812 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a1b400e-f3c1-456c-9d7b-0639a86edbd5" containerName="registry" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.134891 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a1d0b1c-9bae-483f-886d-bf4acffee7d9" containerName="route-controller-manager" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.134903 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a1b400e-f3c1-456c-9d7b-0639a86edbd5" containerName="registry" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.134911 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2ee8da6-43e0-496d-aea1-2c7849305419" containerName="controller-manager" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.134918 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="d777f4d5-fab6-4033-9200-e7cce416dab9" containerName="collect-profiles" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.135263 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.137886 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.138429 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.138697 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.138948 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.139308 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.139543 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.176816 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p"] Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.191411 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh"] Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.192045 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.194153 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.194403 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.194539 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.194714 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.194962 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.196034 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.203700 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.207077 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh"] Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.279722 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/842bf09d-1858-4fe2-b82a-6a0454d86acb-serving-cert\") pod \"controller-manager-57c8cb4f4f-r7ldh\" (UID: \"842bf09d-1858-4fe2-b82a-6a0454d86acb\") " pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.279767 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/842bf09d-1858-4fe2-b82a-6a0454d86acb-client-ca\") pod \"controller-manager-57c8cb4f4f-r7ldh\" (UID: \"842bf09d-1858-4fe2-b82a-6a0454d86acb\") " pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.279795 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fab7ba3a-3025-4c8b-9ca3-ddd59df34d06-config\") pod \"route-controller-manager-65779f6c85-2js4p\" (UID: \"fab7ba3a-3025-4c8b-9ca3-ddd59df34d06\") " pod="openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.279831 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2j9f\" (UniqueName: \"kubernetes.io/projected/fab7ba3a-3025-4c8b-9ca3-ddd59df34d06-kube-api-access-k2j9f\") pod \"route-controller-manager-65779f6c85-2js4p\" (UID: \"fab7ba3a-3025-4c8b-9ca3-ddd59df34d06\") " pod="openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.279855 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fab7ba3a-3025-4c8b-9ca3-ddd59df34d06-serving-cert\") pod \"route-controller-manager-65779f6c85-2js4p\" (UID: \"fab7ba3a-3025-4c8b-9ca3-ddd59df34d06\") " pod="openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.279872 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8sx9\" (UniqueName: \"kubernetes.io/projected/842bf09d-1858-4fe2-b82a-6a0454d86acb-kube-api-access-s8sx9\") pod \"controller-manager-57c8cb4f4f-r7ldh\" (UID: \"842bf09d-1858-4fe2-b82a-6a0454d86acb\") " pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.279894 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/842bf09d-1858-4fe2-b82a-6a0454d86acb-proxy-ca-bundles\") pod \"controller-manager-57c8cb4f4f-r7ldh\" (UID: \"842bf09d-1858-4fe2-b82a-6a0454d86acb\") " pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.279910 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/842bf09d-1858-4fe2-b82a-6a0454d86acb-config\") pod \"controller-manager-57c8cb4f4f-r7ldh\" (UID: \"842bf09d-1858-4fe2-b82a-6a0454d86acb\") " pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.279932 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fab7ba3a-3025-4c8b-9ca3-ddd59df34d06-client-ca\") pod \"route-controller-manager-65779f6c85-2js4p\" (UID: \"fab7ba3a-3025-4c8b-9ca3-ddd59df34d06\") " pod="openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.381003 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2j9f\" (UniqueName: \"kubernetes.io/projected/fab7ba3a-3025-4c8b-9ca3-ddd59df34d06-kube-api-access-k2j9f\") pod \"route-controller-manager-65779f6c85-2js4p\" (UID: \"fab7ba3a-3025-4c8b-9ca3-ddd59df34d06\") " pod="openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.381087 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fab7ba3a-3025-4c8b-9ca3-ddd59df34d06-serving-cert\") pod \"route-controller-manager-65779f6c85-2js4p\" (UID: \"fab7ba3a-3025-4c8b-9ca3-ddd59df34d06\") " pod="openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.381130 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8sx9\" (UniqueName: \"kubernetes.io/projected/842bf09d-1858-4fe2-b82a-6a0454d86acb-kube-api-access-s8sx9\") pod \"controller-manager-57c8cb4f4f-r7ldh\" (UID: \"842bf09d-1858-4fe2-b82a-6a0454d86acb\") " pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.381167 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/842bf09d-1858-4fe2-b82a-6a0454d86acb-proxy-ca-bundles\") pod \"controller-manager-57c8cb4f4f-r7ldh\" (UID: \"842bf09d-1858-4fe2-b82a-6a0454d86acb\") " pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.381214 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/842bf09d-1858-4fe2-b82a-6a0454d86acb-config\") pod \"controller-manager-57c8cb4f4f-r7ldh\" (UID: \"842bf09d-1858-4fe2-b82a-6a0454d86acb\") " pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.381243 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fab7ba3a-3025-4c8b-9ca3-ddd59df34d06-client-ca\") pod \"route-controller-manager-65779f6c85-2js4p\" (UID: \"fab7ba3a-3025-4c8b-9ca3-ddd59df34d06\") " pod="openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.381295 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/842bf09d-1858-4fe2-b82a-6a0454d86acb-serving-cert\") pod \"controller-manager-57c8cb4f4f-r7ldh\" (UID: \"842bf09d-1858-4fe2-b82a-6a0454d86acb\") " pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.381316 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/842bf09d-1858-4fe2-b82a-6a0454d86acb-client-ca\") pod \"controller-manager-57c8cb4f4f-r7ldh\" (UID: \"842bf09d-1858-4fe2-b82a-6a0454d86acb\") " pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.381369 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fab7ba3a-3025-4c8b-9ca3-ddd59df34d06-config\") pod \"route-controller-manager-65779f6c85-2js4p\" (UID: \"fab7ba3a-3025-4c8b-9ca3-ddd59df34d06\") " pod="openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.382747 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/842bf09d-1858-4fe2-b82a-6a0454d86acb-client-ca\") pod \"controller-manager-57c8cb4f4f-r7ldh\" (UID: \"842bf09d-1858-4fe2-b82a-6a0454d86acb\") " pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.382781 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/842bf09d-1858-4fe2-b82a-6a0454d86acb-proxy-ca-bundles\") pod \"controller-manager-57c8cb4f4f-r7ldh\" (UID: \"842bf09d-1858-4fe2-b82a-6a0454d86acb\") " pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.382887 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fab7ba3a-3025-4c8b-9ca3-ddd59df34d06-client-ca\") pod \"route-controller-manager-65779f6c85-2js4p\" (UID: \"fab7ba3a-3025-4c8b-9ca3-ddd59df34d06\") " pod="openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.383024 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/842bf09d-1858-4fe2-b82a-6a0454d86acb-config\") pod \"controller-manager-57c8cb4f4f-r7ldh\" (UID: \"842bf09d-1858-4fe2-b82a-6a0454d86acb\") " pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.383139 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fab7ba3a-3025-4c8b-9ca3-ddd59df34d06-config\") pod \"route-controller-manager-65779f6c85-2js4p\" (UID: \"fab7ba3a-3025-4c8b-9ca3-ddd59df34d06\") " pod="openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.385352 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fab7ba3a-3025-4c8b-9ca3-ddd59df34d06-serving-cert\") pod \"route-controller-manager-65779f6c85-2js4p\" (UID: \"fab7ba3a-3025-4c8b-9ca3-ddd59df34d06\") " pod="openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.385354 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/842bf09d-1858-4fe2-b82a-6a0454d86acb-serving-cert\") pod \"controller-manager-57c8cb4f4f-r7ldh\" (UID: \"842bf09d-1858-4fe2-b82a-6a0454d86acb\") " pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.396745 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2j9f\" (UniqueName: \"kubernetes.io/projected/fab7ba3a-3025-4c8b-9ca3-ddd59df34d06-kube-api-access-k2j9f\") pod \"route-controller-manager-65779f6c85-2js4p\" (UID: \"fab7ba3a-3025-4c8b-9ca3-ddd59df34d06\") " pod="openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.398394 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8sx9\" (UniqueName: \"kubernetes.io/projected/842bf09d-1858-4fe2-b82a-6a0454d86acb-kube-api-access-s8sx9\") pod \"controller-manager-57c8cb4f4f-r7ldh\" (UID: \"842bf09d-1858-4fe2-b82a-6a0454d86acb\") " pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.450324 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.506785 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.764482 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p"] Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.825394 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh"] Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.921770 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a1d0b1c-9bae-483f-886d-bf4acffee7d9" path="/var/lib/kubelet/pods/5a1d0b1c-9bae-483f-886d-bf4acffee7d9/volumes" Oct 03 13:03:48 crc kubenswrapper[4578]: I1003 13:03:48.922816 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2ee8da6-43e0-496d-aea1-2c7849305419" path="/var/lib/kubelet/pods/b2ee8da6-43e0-496d-aea1-2c7849305419/volumes" Oct 03 13:03:49 crc kubenswrapper[4578]: I1003 13:03:49.339067 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" event={"ID":"842bf09d-1858-4fe2-b82a-6a0454d86acb","Type":"ContainerStarted","Data":"f74be1d1e698de28a77be900bf8061d1b25246104b1ac8816654437ada84cf3a"} Oct 03 13:03:49 crc kubenswrapper[4578]: I1003 13:03:49.339416 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" event={"ID":"842bf09d-1858-4fe2-b82a-6a0454d86acb","Type":"ContainerStarted","Data":"d77c2618bc9c9c9ed44359513b9668d8300133d5fad093a5884cda6fc4957c63"} Oct 03 13:03:49 crc kubenswrapper[4578]: I1003 13:03:49.339438 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" Oct 03 13:03:49 crc kubenswrapper[4578]: I1003 13:03:49.341473 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p" event={"ID":"fab7ba3a-3025-4c8b-9ca3-ddd59df34d06","Type":"ContainerStarted","Data":"6c3d8b1611542b466c544dd464e2a45856bec4f7b91430c4ffc182efcdf536eb"} Oct 03 13:03:49 crc kubenswrapper[4578]: I1003 13:03:49.341516 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p" event={"ID":"fab7ba3a-3025-4c8b-9ca3-ddd59df34d06","Type":"ContainerStarted","Data":"7746c326fba7c7e9c2e55a4c603d6f503d0f654a3865482c9ab5fe66f4de2f89"} Oct 03 13:03:49 crc kubenswrapper[4578]: I1003 13:03:49.341658 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p" Oct 03 13:03:49 crc kubenswrapper[4578]: I1003 13:03:49.356580 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p" Oct 03 13:03:49 crc kubenswrapper[4578]: I1003 13:03:49.358299 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" podStartSLOduration=3.358282805 podStartE2EDuration="3.358282805s" podCreationTimestamp="2025-10-03 13:03:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:03:49.357894103 +0000 UTC m=+765.156366307" watchObservedRunningTime="2025-10-03 13:03:49.358282805 +0000 UTC m=+765.156754989" Oct 03 13:03:49 crc kubenswrapper[4578]: I1003 13:03:49.359542 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-57c8cb4f4f-r7ldh" Oct 03 13:03:49 crc kubenswrapper[4578]: I1003 13:03:49.381416 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-65779f6c85-2js4p" podStartSLOduration=1.381395179 podStartE2EDuration="1.381395179s" podCreationTimestamp="2025-10-03 13:03:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:03:49.379702117 +0000 UTC m=+765.178174301" watchObservedRunningTime="2025-10-03 13:03:49.381395179 +0000 UTC m=+765.179867373" Oct 03 13:03:55 crc kubenswrapper[4578]: I1003 13:03:55.563877 4578 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.054992 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-d4qcd"] Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.056451 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-d4qcd" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.064194 4578 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-b9bln" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.064288 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.070419 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.075014 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-d4qcd"] Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.082852 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-rb7jq"] Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.083476 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-rb7jq" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.085392 4578 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-bv4hr" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.117598 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xp9mm"] Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.118826 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-xp9mm" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.125917 4578 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-rr45r" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.133157 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xp9mm"] Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.152067 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzg6f\" (UniqueName: \"kubernetes.io/projected/fde8ee73-946e-4c94-85ff-34f54a387575-kube-api-access-hzg6f\") pod \"cert-manager-cainjector-7f985d654d-d4qcd\" (UID: \"fde8ee73-946e-4c94-85ff-34f54a387575\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-d4qcd" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.170858 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-rb7jq"] Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.253732 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2ndq\" (UniqueName: \"kubernetes.io/projected/5cb84a0f-4490-4370-b07e-71ecb650fc07-kube-api-access-f2ndq\") pod \"cert-manager-webhook-5655c58dd6-xp9mm\" (UID: \"5cb84a0f-4490-4370-b07e-71ecb650fc07\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xp9mm" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.253783 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzg6f\" (UniqueName: \"kubernetes.io/projected/fde8ee73-946e-4c94-85ff-34f54a387575-kube-api-access-hzg6f\") pod \"cert-manager-cainjector-7f985d654d-d4qcd\" (UID: \"fde8ee73-946e-4c94-85ff-34f54a387575\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-d4qcd" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.253833 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msfnc\" (UniqueName: \"kubernetes.io/projected/7350f95d-fefa-47f4-85fb-9ac7a9753bec-kube-api-access-msfnc\") pod \"cert-manager-5b446d88c5-rb7jq\" (UID: \"7350f95d-fefa-47f4-85fb-9ac7a9753bec\") " pod="cert-manager/cert-manager-5b446d88c5-rb7jq" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.274691 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzg6f\" (UniqueName: \"kubernetes.io/projected/fde8ee73-946e-4c94-85ff-34f54a387575-kube-api-access-hzg6f\") pod \"cert-manager-cainjector-7f985d654d-d4qcd\" (UID: \"fde8ee73-946e-4c94-85ff-34f54a387575\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-d4qcd" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.355035 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2ndq\" (UniqueName: \"kubernetes.io/projected/5cb84a0f-4490-4370-b07e-71ecb650fc07-kube-api-access-f2ndq\") pod \"cert-manager-webhook-5655c58dd6-xp9mm\" (UID: \"5cb84a0f-4490-4370-b07e-71ecb650fc07\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xp9mm" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.355113 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msfnc\" (UniqueName: \"kubernetes.io/projected/7350f95d-fefa-47f4-85fb-9ac7a9753bec-kube-api-access-msfnc\") pod \"cert-manager-5b446d88c5-rb7jq\" (UID: \"7350f95d-fefa-47f4-85fb-9ac7a9753bec\") " pod="cert-manager/cert-manager-5b446d88c5-rb7jq" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.371556 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-d4qcd" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.373529 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msfnc\" (UniqueName: \"kubernetes.io/projected/7350f95d-fefa-47f4-85fb-9ac7a9753bec-kube-api-access-msfnc\") pod \"cert-manager-5b446d88c5-rb7jq\" (UID: \"7350f95d-fefa-47f4-85fb-9ac7a9753bec\") " pod="cert-manager/cert-manager-5b446d88c5-rb7jq" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.373708 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2ndq\" (UniqueName: \"kubernetes.io/projected/5cb84a0f-4490-4370-b07e-71ecb650fc07-kube-api-access-f2ndq\") pod \"cert-manager-webhook-5655c58dd6-xp9mm\" (UID: \"5cb84a0f-4490-4370-b07e-71ecb650fc07\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xp9mm" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.396282 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-rb7jq" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.432247 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-xp9mm" Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.813622 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-d4qcd"] Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.819978 4578 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.894401 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-rb7jq"] Oct 03 13:04:10 crc kubenswrapper[4578]: W1003 13:04:10.897653 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7350f95d_fefa_47f4_85fb_9ac7a9753bec.slice/crio-dea02d73ba37eee59867d934a426e1a5bc16ab5ae8de26190f7810e3b7785cc5 WatchSource:0}: Error finding container dea02d73ba37eee59867d934a426e1a5bc16ab5ae8de26190f7810e3b7785cc5: Status 404 returned error can't find the container with id dea02d73ba37eee59867d934a426e1a5bc16ab5ae8de26190f7810e3b7785cc5 Oct 03 13:04:10 crc kubenswrapper[4578]: I1003 13:04:10.964284 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xp9mm"] Oct 03 13:04:11 crc kubenswrapper[4578]: I1003 13:04:11.458534 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-rb7jq" event={"ID":"7350f95d-fefa-47f4-85fb-9ac7a9753bec","Type":"ContainerStarted","Data":"dea02d73ba37eee59867d934a426e1a5bc16ab5ae8de26190f7810e3b7785cc5"} Oct 03 13:04:11 crc kubenswrapper[4578]: I1003 13:04:11.459597 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-d4qcd" event={"ID":"fde8ee73-946e-4c94-85ff-34f54a387575","Type":"ContainerStarted","Data":"e471b410a2cd5fb7e0a0e595d161249ade636edc853aec9c09d39633cbb3d90e"} Oct 03 13:04:11 crc kubenswrapper[4578]: I1003 13:04:11.460657 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-xp9mm" event={"ID":"5cb84a0f-4490-4370-b07e-71ecb650fc07","Type":"ContainerStarted","Data":"e7cba15995765ae9674d8b8cfe4412d40518ce0f5193c58ae34c76c91bb466b4"} Oct 03 13:04:15 crc kubenswrapper[4578]: I1003 13:04:15.483218 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-rb7jq" event={"ID":"7350f95d-fefa-47f4-85fb-9ac7a9753bec","Type":"ContainerStarted","Data":"bd87d8c7446781fa57c90808bc8aa06e0d46de7c84c5679954c4e24af3cbdba2"} Oct 03 13:04:15 crc kubenswrapper[4578]: I1003 13:04:15.484848 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-d4qcd" event={"ID":"fde8ee73-946e-4c94-85ff-34f54a387575","Type":"ContainerStarted","Data":"434d78e2b48f188225c8ac26220ec2ab7f42def90c86263a40b1ff130c315fc5"} Oct 03 13:04:15 crc kubenswrapper[4578]: I1003 13:04:15.499431 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-rb7jq" podStartSLOduration=1.265583908 podStartE2EDuration="5.499409923s" podCreationTimestamp="2025-10-03 13:04:10 +0000 UTC" firstStartedPulling="2025-10-03 13:04:10.900215629 +0000 UTC m=+786.698687813" lastFinishedPulling="2025-10-03 13:04:15.134041644 +0000 UTC m=+790.932513828" observedRunningTime="2025-10-03 13:04:15.496602979 +0000 UTC m=+791.295075173" watchObservedRunningTime="2025-10-03 13:04:15.499409923 +0000 UTC m=+791.297882107" Oct 03 13:04:16 crc kubenswrapper[4578]: I1003 13:04:16.491393 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-xp9mm" event={"ID":"5cb84a0f-4490-4370-b07e-71ecb650fc07","Type":"ContainerStarted","Data":"a0a14c69ca44728d379af4c6955c41ea7c5c0ea2b2ce87d44e5ad46572faea2b"} Oct 03 13:04:16 crc kubenswrapper[4578]: I1003 13:04:16.508427 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-d4qcd" podStartSLOduration=2.2709140039999998 podStartE2EDuration="6.50840416s" podCreationTimestamp="2025-10-03 13:04:10 +0000 UTC" firstStartedPulling="2025-10-03 13:04:10.819560702 +0000 UTC m=+786.618032886" lastFinishedPulling="2025-10-03 13:04:15.057050868 +0000 UTC m=+790.855523042" observedRunningTime="2025-10-03 13:04:15.52344438 +0000 UTC m=+791.321916584" watchObservedRunningTime="2025-10-03 13:04:16.50840416 +0000 UTC m=+792.306876354" Oct 03 13:04:16 crc kubenswrapper[4578]: I1003 13:04:16.508820 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-xp9mm" podStartSLOduration=1.308598279 podStartE2EDuration="6.508812532s" podCreationTimestamp="2025-10-03 13:04:10 +0000 UTC" firstStartedPulling="2025-10-03 13:04:10.983190647 +0000 UTC m=+786.781662841" lastFinishedPulling="2025-10-03 13:04:16.18340491 +0000 UTC m=+791.981877094" observedRunningTime="2025-10-03 13:04:16.504947636 +0000 UTC m=+792.303419810" watchObservedRunningTime="2025-10-03 13:04:16.508812532 +0000 UTC m=+792.307284736" Oct 03 13:04:17 crc kubenswrapper[4578]: I1003 13:04:17.497702 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-xp9mm" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.512800 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8qvh9"] Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.513504 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovn-controller" containerID="cri-o://73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030" gracePeriod=30 Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.513592 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="nbdb" containerID="cri-o://4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81" gracePeriod=30 Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.513626 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236" gracePeriod=30 Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.513691 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="kube-rbac-proxy-node" containerID="cri-o://31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d" gracePeriod=30 Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.513732 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovn-acl-logging" containerID="cri-o://39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9" gracePeriod=30 Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.513610 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="northd" containerID="cri-o://c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982" gracePeriod=30 Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.514000 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="sbdb" containerID="cri-o://377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40" gracePeriod=30 Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.548531 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovnkube-controller" containerID="cri-o://ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01" gracePeriod=30 Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.856559 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8qvh9_ebaa96c5-be7c-4f05-819d-f37e1ceddac8/ovnkube-controller/3.log" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.858764 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8qvh9_ebaa96c5-be7c-4f05-819d-f37e1ceddac8/ovn-acl-logging/0.log" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.859220 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8qvh9_ebaa96c5-be7c-4f05-819d-f37e1ceddac8/ovn-controller/0.log" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.859690 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.922764 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-z4fwm"] Oct 03 13:04:20 crc kubenswrapper[4578]: E1003 13:04:20.922949 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="kube-rbac-proxy-node" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.922959 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="kube-rbac-proxy-node" Oct 03 13:04:20 crc kubenswrapper[4578]: E1003 13:04:20.922970 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovn-acl-logging" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.922976 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovn-acl-logging" Oct 03 13:04:20 crc kubenswrapper[4578]: E1003 13:04:20.922984 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovnkube-controller" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.922990 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovnkube-controller" Oct 03 13:04:20 crc kubenswrapper[4578]: E1003 13:04:20.922999 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovn-controller" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923012 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovn-controller" Oct 03 13:04:20 crc kubenswrapper[4578]: E1003 13:04:20.923024 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="kubecfg-setup" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923030 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="kubecfg-setup" Oct 03 13:04:20 crc kubenswrapper[4578]: E1003 13:04:20.923043 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="nbdb" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923048 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="nbdb" Oct 03 13:04:20 crc kubenswrapper[4578]: E1003 13:04:20.923058 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="sbdb" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923063 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="sbdb" Oct 03 13:04:20 crc kubenswrapper[4578]: E1003 13:04:20.923069 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovnkube-controller" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923075 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovnkube-controller" Oct 03 13:04:20 crc kubenswrapper[4578]: E1003 13:04:20.923081 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovnkube-controller" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923087 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovnkube-controller" Oct 03 13:04:20 crc kubenswrapper[4578]: E1003 13:04:20.923096 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="northd" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923103 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="northd" Oct 03 13:04:20 crc kubenswrapper[4578]: E1003 13:04:20.923112 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="kube-rbac-proxy-ovn-metrics" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923121 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="kube-rbac-proxy-ovn-metrics" Oct 03 13:04:20 crc kubenswrapper[4578]: E1003 13:04:20.923133 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovnkube-controller" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923138 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovnkube-controller" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923234 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="nbdb" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923248 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="kube-rbac-proxy-node" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923257 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="sbdb" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923270 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovnkube-controller" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923280 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovnkube-controller" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923288 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="kube-rbac-proxy-ovn-metrics" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923299 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovn-controller" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923308 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovnkube-controller" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923314 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovnkube-controller" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923325 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="northd" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923331 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovn-acl-logging" Oct 03 13:04:20 crc kubenswrapper[4578]: E1003 13:04:20.923434 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovnkube-controller" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923443 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovnkube-controller" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.923543 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerName="ovnkube-controller" Oct 03 13:04:20 crc kubenswrapper[4578]: I1003 13:04:20.925142 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.003646 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-run-openvswitch\") pod \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.003720 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-env-overrides\") pod \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.003725 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "ebaa96c5-be7c-4f05-819d-f37e1ceddac8" (UID: "ebaa96c5-be7c-4f05-819d-f37e1ceddac8"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.003758 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gt6c5\" (UniqueName: \"kubernetes.io/projected/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-kube-api-access-gt6c5\") pod \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.003795 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-etc-openvswitch\") pod \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.003813 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-ovnkube-config\") pod \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.003836 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.003871 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "ebaa96c5-be7c-4f05-819d-f37e1ceddac8" (UID: "ebaa96c5-be7c-4f05-819d-f37e1ceddac8"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.003968 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-cni-netd\") pod \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004005 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "ebaa96c5-be7c-4f05-819d-f37e1ceddac8" (UID: "ebaa96c5-be7c-4f05-819d-f37e1ceddac8"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004172 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "ebaa96c5-be7c-4f05-819d-f37e1ceddac8" (UID: "ebaa96c5-be7c-4f05-819d-f37e1ceddac8"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004264 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "ebaa96c5-be7c-4f05-819d-f37e1ceddac8" (UID: "ebaa96c5-be7c-4f05-819d-f37e1ceddac8"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004320 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-run-systemd\") pod \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004355 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-cni-bin\") pod \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004373 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-run-netns\") pod \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004394 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-run-ovn-kubernetes\") pod \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004419 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-node-log\") pod \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004437 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-slash\") pod \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004457 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-run-ovn\") pod \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004467 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "ebaa96c5-be7c-4f05-819d-f37e1ceddac8" (UID: "ebaa96c5-be7c-4f05-819d-f37e1ceddac8"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004478 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-kubelet\") pod \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004502 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "ebaa96c5-be7c-4f05-819d-f37e1ceddac8" (UID: "ebaa96c5-be7c-4f05-819d-f37e1ceddac8"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004511 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-systemd-units\") pod \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004533 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-var-lib-openvswitch\") pod \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004560 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-ovn-node-metrics-cert\") pod \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004586 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-ovnkube-script-lib\") pod \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004608 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-log-socket\") pod \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\" (UID: \"ebaa96c5-be7c-4f05-819d-f37e1ceddac8\") " Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004533 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "ebaa96c5-be7c-4f05-819d-f37e1ceddac8" (UID: "ebaa96c5-be7c-4f05-819d-f37e1ceddac8"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004528 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "ebaa96c5-be7c-4f05-819d-f37e1ceddac8" (UID: "ebaa96c5-be7c-4f05-819d-f37e1ceddac8"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004553 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "ebaa96c5-be7c-4f05-819d-f37e1ceddac8" (UID: "ebaa96c5-be7c-4f05-819d-f37e1ceddac8"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004558 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-slash" (OuterVolumeSpecName: "host-slash") pod "ebaa96c5-be7c-4f05-819d-f37e1ceddac8" (UID: "ebaa96c5-be7c-4f05-819d-f37e1ceddac8"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004577 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "ebaa96c5-be7c-4f05-819d-f37e1ceddac8" (UID: "ebaa96c5-be7c-4f05-819d-f37e1ceddac8"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004579 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-node-log" (OuterVolumeSpecName: "node-log") pod "ebaa96c5-be7c-4f05-819d-f37e1ceddac8" (UID: "ebaa96c5-be7c-4f05-819d-f37e1ceddac8"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004599 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "ebaa96c5-be7c-4f05-819d-f37e1ceddac8" (UID: "ebaa96c5-be7c-4f05-819d-f37e1ceddac8"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004604 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "ebaa96c5-be7c-4f05-819d-f37e1ceddac8" (UID: "ebaa96c5-be7c-4f05-819d-f37e1ceddac8"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004746 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004778 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-log-socket" (OuterVolumeSpecName: "log-socket") pod "ebaa96c5-be7c-4f05-819d-f37e1ceddac8" (UID: "ebaa96c5-be7c-4f05-819d-f37e1ceddac8"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004788 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9761d56d-159d-4161-9ecc-13165e7b758d-ovnkube-config\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004853 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-var-lib-openvswitch\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004886 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjzmp\" (UniqueName: \"kubernetes.io/projected/9761d56d-159d-4161-9ecc-13165e7b758d-kube-api-access-sjzmp\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004914 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "ebaa96c5-be7c-4f05-819d-f37e1ceddac8" (UID: "ebaa96c5-be7c-4f05-819d-f37e1ceddac8"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004923 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-slash\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.004977 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-run-ovn-kubernetes\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005018 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-cni-netd\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005036 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9761d56d-159d-4161-9ecc-13165e7b758d-ovnkube-script-lib\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005077 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-kubelet\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005095 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-run-ovn\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005125 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-run-netns\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005148 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9761d56d-159d-4161-9ecc-13165e7b758d-ovn-node-metrics-cert\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005168 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-node-log\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005187 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-systemd-units\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005212 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-etc-openvswitch\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005291 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-log-socket\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005320 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9761d56d-159d-4161-9ecc-13165e7b758d-env-overrides\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005538 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-run-systemd\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005556 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-cni-bin\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005608 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-run-openvswitch\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005673 4578 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005687 4578 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005704 4578 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005715 4578 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005727 4578 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005749 4578 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005761 4578 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005771 4578 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005782 4578 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005793 4578 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-node-log\") on node \"crc\" DevicePath \"\"" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005804 4578 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-slash\") on node \"crc\" DevicePath \"\"" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005816 4578 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005827 4578 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005837 4578 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005848 4578 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005858 4578 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.005868 4578 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-log-socket\") on node \"crc\" DevicePath \"\"" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.009402 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "ebaa96c5-be7c-4f05-819d-f37e1ceddac8" (UID: "ebaa96c5-be7c-4f05-819d-f37e1ceddac8"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.009671 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-kube-api-access-gt6c5" (OuterVolumeSpecName: "kube-api-access-gt6c5") pod "ebaa96c5-be7c-4f05-819d-f37e1ceddac8" (UID: "ebaa96c5-be7c-4f05-819d-f37e1ceddac8"). InnerVolumeSpecName "kube-api-access-gt6c5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.017789 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "ebaa96c5-be7c-4f05-819d-f37e1ceddac8" (UID: "ebaa96c5-be7c-4f05-819d-f37e1ceddac8"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.106673 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-run-openvswitch\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.106720 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.106745 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9761d56d-159d-4161-9ecc-13165e7b758d-ovnkube-config\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.106796 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-var-lib-openvswitch\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.106816 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjzmp\" (UniqueName: \"kubernetes.io/projected/9761d56d-159d-4161-9ecc-13165e7b758d-kube-api-access-sjzmp\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.106816 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-run-openvswitch\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.106830 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-slash\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.106817 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.106851 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-run-ovn-kubernetes\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.106872 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-run-ovn-kubernetes\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.106875 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-var-lib-openvswitch\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.106897 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-cni-netd\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.106912 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9761d56d-159d-4161-9ecc-13165e7b758d-ovnkube-script-lib\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.106929 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-run-ovn\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.106942 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-kubelet\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.106958 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-run-netns\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.106972 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9761d56d-159d-4161-9ecc-13165e7b758d-ovn-node-metrics-cert\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.106986 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-node-log\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.107001 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-etc-openvswitch\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.107014 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-systemd-units\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.107034 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-log-socket\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.107048 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9761d56d-159d-4161-9ecc-13165e7b758d-env-overrides\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.107069 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-run-systemd\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.107084 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-cni-bin\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.107124 4578 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.108673 4578 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.108691 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gt6c5\" (UniqueName: \"kubernetes.io/projected/ebaa96c5-be7c-4f05-819d-f37e1ceddac8-kube-api-access-gt6c5\") on node \"crc\" DevicePath \"\"" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.107601 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-node-log\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.107655 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-etc-openvswitch\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.107680 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-log-socket\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.107702 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-run-netns\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.107576 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-systemd-units\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.107723 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-slash\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.107719 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-kubelet\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.107736 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-run-ovn\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.107743 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-cni-netd\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.107749 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-host-cni-bin\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.107967 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9761d56d-159d-4161-9ecc-13165e7b758d-env-overrides\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.108235 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9761d56d-159d-4161-9ecc-13165e7b758d-ovnkube-script-lib\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.107696 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9761d56d-159d-4161-9ecc-13165e7b758d-run-systemd\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.107582 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9761d56d-159d-4161-9ecc-13165e7b758d-ovnkube-config\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.110687 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9761d56d-159d-4161-9ecc-13165e7b758d-ovn-node-metrics-cert\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.123866 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjzmp\" (UniqueName: \"kubernetes.io/projected/9761d56d-159d-4161-9ecc-13165e7b758d-kube-api-access-sjzmp\") pod \"ovnkube-node-z4fwm\" (UID: \"9761d56d-159d-4161-9ecc-13165e7b758d\") " pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.238057 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.520574 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-c4jgn_fed3968f-4bb0-4209-a377-16c26f9ea070/kube-multus/2.log" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.521673 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-c4jgn_fed3968f-4bb0-4209-a377-16c26f9ea070/kube-multus/1.log" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.521715 4578 generic.go:334] "Generic (PLEG): container finished" podID="fed3968f-4bb0-4209-a377-16c26f9ea070" containerID="8863d0d89e062a4984f279b5b3f315ff53e3502946c377ef60a2481ab6294084" exitCode=2 Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.521812 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-c4jgn" event={"ID":"fed3968f-4bb0-4209-a377-16c26f9ea070","Type":"ContainerDied","Data":"8863d0d89e062a4984f279b5b3f315ff53e3502946c377ef60a2481ab6294084"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.521898 4578 scope.go:117] "RemoveContainer" containerID="650520feb6cfcf5c6aea04b4a990593eab6f2a2ddec00e2270105f25ee9a067d" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.522322 4578 scope.go:117] "RemoveContainer" containerID="8863d0d89e062a4984f279b5b3f315ff53e3502946c377ef60a2481ab6294084" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.526504 4578 generic.go:334] "Generic (PLEG): container finished" podID="9761d56d-159d-4161-9ecc-13165e7b758d" containerID="805f30c23f67f58ba522927de7cfbb1ec6870a4376f2841a4c0dfbe7363f08b9" exitCode=0 Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.526654 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" event={"ID":"9761d56d-159d-4161-9ecc-13165e7b758d","Type":"ContainerDied","Data":"805f30c23f67f58ba522927de7cfbb1ec6870a4376f2841a4c0dfbe7363f08b9"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.526731 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" event={"ID":"9761d56d-159d-4161-9ecc-13165e7b758d","Type":"ContainerStarted","Data":"8aa9797fab5291443d8549e9822b5c6cab93d7af037928f414900336cc6641e2"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.582161 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8qvh9_ebaa96c5-be7c-4f05-819d-f37e1ceddac8/ovnkube-controller/3.log" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.585953 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8qvh9_ebaa96c5-be7c-4f05-819d-f37e1ceddac8/ovn-acl-logging/0.log" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.586549 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8qvh9_ebaa96c5-be7c-4f05-819d-f37e1ceddac8/ovn-controller/0.log" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589437 4578 generic.go:334] "Generic (PLEG): container finished" podID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerID="ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01" exitCode=0 Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589463 4578 generic.go:334] "Generic (PLEG): container finished" podID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerID="377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40" exitCode=0 Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589470 4578 generic.go:334] "Generic (PLEG): container finished" podID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerID="4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81" exitCode=0 Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589476 4578 generic.go:334] "Generic (PLEG): container finished" podID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerID="c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982" exitCode=0 Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589484 4578 generic.go:334] "Generic (PLEG): container finished" podID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerID="32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236" exitCode=0 Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589490 4578 generic.go:334] "Generic (PLEG): container finished" podID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerID="31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d" exitCode=0 Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589497 4578 generic.go:334] "Generic (PLEG): container finished" podID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerID="39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9" exitCode=143 Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589504 4578 generic.go:334] "Generic (PLEG): container finished" podID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" containerID="73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030" exitCode=143 Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589523 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerDied","Data":"ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589548 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerDied","Data":"377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589559 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerDied","Data":"4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589567 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerDied","Data":"c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589577 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerDied","Data":"32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589585 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerDied","Data":"31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589594 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589603 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589608 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589613 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589619 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589623 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589646 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589651 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589656 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589660 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589667 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerDied","Data":"39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589675 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589680 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589686 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589693 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589698 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589702 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589707 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589713 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589718 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589723 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589729 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerDied","Data":"73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589736 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589741 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589746 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589752 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589757 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589762 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589766 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589771 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589776 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589781 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589788 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" event={"ID":"ebaa96c5-be7c-4f05-819d-f37e1ceddac8","Type":"ContainerDied","Data":"cac09449164b0eee0488dbd39377414aa0d07f6a27a355af9d775b75b6e18563"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589794 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589800 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589806 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589813 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589820 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589826 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589832 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589839 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589846 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589851 4578 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49"} Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.589953 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8qvh9" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.694556 4578 scope.go:117] "RemoveContainer" containerID="ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.707018 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8qvh9"] Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.714369 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8qvh9"] Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.716588 4578 scope.go:117] "RemoveContainer" containerID="508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.740771 4578 scope.go:117] "RemoveContainer" containerID="377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.766158 4578 scope.go:117] "RemoveContainer" containerID="4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.784929 4578 scope.go:117] "RemoveContainer" containerID="c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.798253 4578 scope.go:117] "RemoveContainer" containerID="32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.810218 4578 scope.go:117] "RemoveContainer" containerID="31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.822315 4578 scope.go:117] "RemoveContainer" containerID="39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.837485 4578 scope.go:117] "RemoveContainer" containerID="73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.849755 4578 scope.go:117] "RemoveContainer" containerID="0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.863985 4578 scope.go:117] "RemoveContainer" containerID="ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01" Oct 03 13:04:21 crc kubenswrapper[4578]: E1003 13:04:21.864372 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01\": container with ID starting with ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01 not found: ID does not exist" containerID="ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.864425 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01"} err="failed to get container status \"ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01\": rpc error: code = NotFound desc = could not find container \"ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01\": container with ID starting with ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.864463 4578 scope.go:117] "RemoveContainer" containerID="508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c" Oct 03 13:04:21 crc kubenswrapper[4578]: E1003 13:04:21.864913 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c\": container with ID starting with 508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c not found: ID does not exist" containerID="508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.864943 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c"} err="failed to get container status \"508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c\": rpc error: code = NotFound desc = could not find container \"508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c\": container with ID starting with 508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.864962 4578 scope.go:117] "RemoveContainer" containerID="377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40" Oct 03 13:04:21 crc kubenswrapper[4578]: E1003 13:04:21.865282 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\": container with ID starting with 377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40 not found: ID does not exist" containerID="377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.865313 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40"} err="failed to get container status \"377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\": rpc error: code = NotFound desc = could not find container \"377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\": container with ID starting with 377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.865333 4578 scope.go:117] "RemoveContainer" containerID="4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81" Oct 03 13:04:21 crc kubenswrapper[4578]: E1003 13:04:21.865560 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\": container with ID starting with 4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81 not found: ID does not exist" containerID="4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.865585 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81"} err="failed to get container status \"4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\": rpc error: code = NotFound desc = could not find container \"4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\": container with ID starting with 4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.865603 4578 scope.go:117] "RemoveContainer" containerID="c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982" Oct 03 13:04:21 crc kubenswrapper[4578]: E1003 13:04:21.866010 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\": container with ID starting with c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982 not found: ID does not exist" containerID="c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.866052 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982"} err="failed to get container status \"c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\": rpc error: code = NotFound desc = could not find container \"c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\": container with ID starting with c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.866071 4578 scope.go:117] "RemoveContainer" containerID="32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236" Oct 03 13:04:21 crc kubenswrapper[4578]: E1003 13:04:21.866326 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\": container with ID starting with 32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236 not found: ID does not exist" containerID="32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.866353 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236"} err="failed to get container status \"32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\": rpc error: code = NotFound desc = could not find container \"32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\": container with ID starting with 32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.866392 4578 scope.go:117] "RemoveContainer" containerID="31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d" Oct 03 13:04:21 crc kubenswrapper[4578]: E1003 13:04:21.866587 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\": container with ID starting with 31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d not found: ID does not exist" containerID="31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.866612 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d"} err="failed to get container status \"31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\": rpc error: code = NotFound desc = could not find container \"31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\": container with ID starting with 31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.866644 4578 scope.go:117] "RemoveContainer" containerID="39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9" Oct 03 13:04:21 crc kubenswrapper[4578]: E1003 13:04:21.866898 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\": container with ID starting with 39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9 not found: ID does not exist" containerID="39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.866920 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9"} err="failed to get container status \"39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\": rpc error: code = NotFound desc = could not find container \"39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\": container with ID starting with 39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.866936 4578 scope.go:117] "RemoveContainer" containerID="73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030" Oct 03 13:04:21 crc kubenswrapper[4578]: E1003 13:04:21.867155 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\": container with ID starting with 73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030 not found: ID does not exist" containerID="73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.867175 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030"} err="failed to get container status \"73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\": rpc error: code = NotFound desc = could not find container \"73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\": container with ID starting with 73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.867190 4578 scope.go:117] "RemoveContainer" containerID="0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49" Oct 03 13:04:21 crc kubenswrapper[4578]: E1003 13:04:21.867402 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\": container with ID starting with 0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49 not found: ID does not exist" containerID="0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.867430 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49"} err="failed to get container status \"0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\": rpc error: code = NotFound desc = could not find container \"0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\": container with ID starting with 0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.867449 4578 scope.go:117] "RemoveContainer" containerID="ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.867671 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01"} err="failed to get container status \"ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01\": rpc error: code = NotFound desc = could not find container \"ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01\": container with ID starting with ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.867697 4578 scope.go:117] "RemoveContainer" containerID="508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.868857 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c"} err="failed to get container status \"508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c\": rpc error: code = NotFound desc = could not find container \"508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c\": container with ID starting with 508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.868877 4578 scope.go:117] "RemoveContainer" containerID="377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.869379 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40"} err="failed to get container status \"377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\": rpc error: code = NotFound desc = could not find container \"377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\": container with ID starting with 377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.869396 4578 scope.go:117] "RemoveContainer" containerID="4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.869729 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81"} err="failed to get container status \"4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\": rpc error: code = NotFound desc = could not find container \"4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\": container with ID starting with 4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.869747 4578 scope.go:117] "RemoveContainer" containerID="c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.870078 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982"} err="failed to get container status \"c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\": rpc error: code = NotFound desc = could not find container \"c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\": container with ID starting with c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.870117 4578 scope.go:117] "RemoveContainer" containerID="32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.870741 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236"} err="failed to get container status \"32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\": rpc error: code = NotFound desc = could not find container \"32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\": container with ID starting with 32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.870768 4578 scope.go:117] "RemoveContainer" containerID="31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.871133 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d"} err="failed to get container status \"31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\": rpc error: code = NotFound desc = could not find container \"31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\": container with ID starting with 31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.871153 4578 scope.go:117] "RemoveContainer" containerID="39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.871452 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9"} err="failed to get container status \"39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\": rpc error: code = NotFound desc = could not find container \"39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\": container with ID starting with 39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.871474 4578 scope.go:117] "RemoveContainer" containerID="73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.871694 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030"} err="failed to get container status \"73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\": rpc error: code = NotFound desc = could not find container \"73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\": container with ID starting with 73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.871710 4578 scope.go:117] "RemoveContainer" containerID="0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.871912 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49"} err="failed to get container status \"0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\": rpc error: code = NotFound desc = could not find container \"0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\": container with ID starting with 0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.871941 4578 scope.go:117] "RemoveContainer" containerID="ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.872204 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01"} err="failed to get container status \"ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01\": rpc error: code = NotFound desc = could not find container \"ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01\": container with ID starting with ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.872235 4578 scope.go:117] "RemoveContainer" containerID="508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.872594 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c"} err="failed to get container status \"508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c\": rpc error: code = NotFound desc = could not find container \"508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c\": container with ID starting with 508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.872613 4578 scope.go:117] "RemoveContainer" containerID="377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.872832 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40"} err="failed to get container status \"377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\": rpc error: code = NotFound desc = could not find container \"377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\": container with ID starting with 377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.872847 4578 scope.go:117] "RemoveContainer" containerID="4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.873049 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81"} err="failed to get container status \"4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\": rpc error: code = NotFound desc = could not find container \"4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\": container with ID starting with 4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.873064 4578 scope.go:117] "RemoveContainer" containerID="c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.873296 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982"} err="failed to get container status \"c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\": rpc error: code = NotFound desc = could not find container \"c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\": container with ID starting with c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.873311 4578 scope.go:117] "RemoveContainer" containerID="32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.873484 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236"} err="failed to get container status \"32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\": rpc error: code = NotFound desc = could not find container \"32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\": container with ID starting with 32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.873498 4578 scope.go:117] "RemoveContainer" containerID="31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.873719 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d"} err="failed to get container status \"31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\": rpc error: code = NotFound desc = could not find container \"31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\": container with ID starting with 31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.873735 4578 scope.go:117] "RemoveContainer" containerID="39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.873908 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9"} err="failed to get container status \"39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\": rpc error: code = NotFound desc = could not find container \"39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\": container with ID starting with 39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.873927 4578 scope.go:117] "RemoveContainer" containerID="73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.874101 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030"} err="failed to get container status \"73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\": rpc error: code = NotFound desc = could not find container \"73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\": container with ID starting with 73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.874119 4578 scope.go:117] "RemoveContainer" containerID="0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.874280 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49"} err="failed to get container status \"0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\": rpc error: code = NotFound desc = could not find container \"0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\": container with ID starting with 0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.874293 4578 scope.go:117] "RemoveContainer" containerID="ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.874591 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01"} err="failed to get container status \"ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01\": rpc error: code = NotFound desc = could not find container \"ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01\": container with ID starting with ec6178ee217efb896d72cd6bd7e8b95a451dc932afce1ffee7ee4e061c708a01 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.874607 4578 scope.go:117] "RemoveContainer" containerID="508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.874906 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c"} err="failed to get container status \"508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c\": rpc error: code = NotFound desc = could not find container \"508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c\": container with ID starting with 508eebbf96d1f4ad9a09be990eaca4bbbde070863a50512665a4e31a39fd0d2c not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.874920 4578 scope.go:117] "RemoveContainer" containerID="377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.875155 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40"} err="failed to get container status \"377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\": rpc error: code = NotFound desc = could not find container \"377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40\": container with ID starting with 377be7d38d6baca97b9e2abcc13e75f2023cab2ec05cb3029302d0ac87e43b40 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.875175 4578 scope.go:117] "RemoveContainer" containerID="4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.875364 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81"} err="failed to get container status \"4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\": rpc error: code = NotFound desc = could not find container \"4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81\": container with ID starting with 4ef2e4357e5cf04b12afa3332d58006d20ffc642403a490caf8186a90dd04b81 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.875379 4578 scope.go:117] "RemoveContainer" containerID="c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.875585 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982"} err="failed to get container status \"c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\": rpc error: code = NotFound desc = could not find container \"c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982\": container with ID starting with c4cfed2f3dc5a0eb6d31cf3f04ebe3152956d916c4c93dea0c6eadd656c1f982 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.875600 4578 scope.go:117] "RemoveContainer" containerID="32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.875795 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236"} err="failed to get container status \"32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\": rpc error: code = NotFound desc = could not find container \"32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236\": container with ID starting with 32a1a0da3294778bd1f07f96fcbcee5d289083c2f0696509ed0f5d30ff0c7236 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.875811 4578 scope.go:117] "RemoveContainer" containerID="31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.875972 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d"} err="failed to get container status \"31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\": rpc error: code = NotFound desc = could not find container \"31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d\": container with ID starting with 31c04e955a9a68cd5ab45b8ba5f7471a7b69fbc1964b0bcaf591d47510d29c2d not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.875987 4578 scope.go:117] "RemoveContainer" containerID="39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.876177 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9"} err="failed to get container status \"39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\": rpc error: code = NotFound desc = could not find container \"39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9\": container with ID starting with 39ccb79a154fcedc22abf11498c3d21d25f761085a451ec2c1e2690f7d2ab6b9 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.876205 4578 scope.go:117] "RemoveContainer" containerID="73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.876379 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030"} err="failed to get container status \"73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\": rpc error: code = NotFound desc = could not find container \"73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030\": container with ID starting with 73b56644e91532641eae58d29305e9475f5d1efb94e290e686d182710957d030 not found: ID does not exist" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.876394 4578 scope.go:117] "RemoveContainer" containerID="0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49" Oct 03 13:04:21 crc kubenswrapper[4578]: I1003 13:04:21.876563 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49"} err="failed to get container status \"0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\": rpc error: code = NotFound desc = could not find container \"0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49\": container with ID starting with 0e2477117e47c4b92ea6d5bf29dc84545971296f0de7e9aca8c0a207e17c7f49 not found: ID does not exist" Oct 03 13:04:22 crc kubenswrapper[4578]: I1003 13:04:22.596707 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-c4jgn_fed3968f-4bb0-4209-a377-16c26f9ea070/kube-multus/2.log" Oct 03 13:04:22 crc kubenswrapper[4578]: I1003 13:04:22.597093 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-c4jgn" event={"ID":"fed3968f-4bb0-4209-a377-16c26f9ea070","Type":"ContainerStarted","Data":"c37dd3590c75ed0e826f0f749edb731f01ab48dd8c2e1ec147a6676b9443b1c0"} Oct 03 13:04:22 crc kubenswrapper[4578]: I1003 13:04:22.601585 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" event={"ID":"9761d56d-159d-4161-9ecc-13165e7b758d","Type":"ContainerStarted","Data":"ba4d7c25dd0db206d8750015a5f40bac28110d593a3ffab0a4946570add4ed2f"} Oct 03 13:04:22 crc kubenswrapper[4578]: I1003 13:04:22.601680 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" event={"ID":"9761d56d-159d-4161-9ecc-13165e7b758d","Type":"ContainerStarted","Data":"badeab46ff96063440fa269e5d03632e41ffcb1a64a01eb40870cf04260f884d"} Oct 03 13:04:22 crc kubenswrapper[4578]: I1003 13:04:22.601696 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" event={"ID":"9761d56d-159d-4161-9ecc-13165e7b758d","Type":"ContainerStarted","Data":"c91faa3fa1f408cb8b302f2b094f54462c7ae32b03dbeb4a2e574b4db2070bd1"} Oct 03 13:04:22 crc kubenswrapper[4578]: I1003 13:04:22.601710 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" event={"ID":"9761d56d-159d-4161-9ecc-13165e7b758d","Type":"ContainerStarted","Data":"7d9710e4cf920d11d906b8e2c68909ba198ab5140bc0d9b362f4df89d93fd5f7"} Oct 03 13:04:22 crc kubenswrapper[4578]: I1003 13:04:22.601719 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" event={"ID":"9761d56d-159d-4161-9ecc-13165e7b758d","Type":"ContainerStarted","Data":"acbea673143eddd638a4b1e92b9fea1c16cdfccf6e95d7644d030ae10ad4a25d"} Oct 03 13:04:22 crc kubenswrapper[4578]: I1003 13:04:22.601727 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" event={"ID":"9761d56d-159d-4161-9ecc-13165e7b758d","Type":"ContainerStarted","Data":"52a39cd1a6ccae8145de8460b2d45108e385e778ce03ccd591072c20c513abe2"} Oct 03 13:04:22 crc kubenswrapper[4578]: I1003 13:04:22.915246 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebaa96c5-be7c-4f05-819d-f37e1ceddac8" path="/var/lib/kubelet/pods/ebaa96c5-be7c-4f05-819d-f37e1ceddac8/volumes" Oct 03 13:04:24 crc kubenswrapper[4578]: I1003 13:04:24.617101 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" event={"ID":"9761d56d-159d-4161-9ecc-13165e7b758d","Type":"ContainerStarted","Data":"69f03c5baaf33d6e683f8948d27489d14f23d788fbf1fb9f7db6f950243d712f"} Oct 03 13:04:25 crc kubenswrapper[4578]: I1003 13:04:25.435827 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-xp9mm" Oct 03 13:04:27 crc kubenswrapper[4578]: I1003 13:04:27.641013 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" event={"ID":"9761d56d-159d-4161-9ecc-13165e7b758d","Type":"ContainerStarted","Data":"4a207ec5d67dbac366d0db80a2d7f3a62a51d87cffa31583a30c27470a443c32"} Oct 03 13:04:27 crc kubenswrapper[4578]: I1003 13:04:27.641735 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:27 crc kubenswrapper[4578]: I1003 13:04:27.641754 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:27 crc kubenswrapper[4578]: I1003 13:04:27.666491 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:27 crc kubenswrapper[4578]: I1003 13:04:27.675001 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" podStartSLOduration=7.674987415 podStartE2EDuration="7.674987415s" podCreationTimestamp="2025-10-03 13:04:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:04:27.671354616 +0000 UTC m=+803.469826800" watchObservedRunningTime="2025-10-03 13:04:27.674987415 +0000 UTC m=+803.473459599" Oct 03 13:04:28 crc kubenswrapper[4578]: I1003 13:04:28.655490 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:28 crc kubenswrapper[4578]: I1003 13:04:28.692048 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:51 crc kubenswrapper[4578]: I1003 13:04:51.276192 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-z4fwm" Oct 03 13:04:56 crc kubenswrapper[4578]: I1003 13:04:56.931533 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bcz4n"] Oct 03 13:04:56 crc kubenswrapper[4578]: I1003 13:04:56.933647 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bcz4n" Oct 03 13:04:56 crc kubenswrapper[4578]: I1003 13:04:56.939364 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bcz4n"] Oct 03 13:04:56 crc kubenswrapper[4578]: I1003 13:04:56.957410 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e97ceac1-f815-4c26-a0eb-34d882101360-catalog-content\") pod \"community-operators-bcz4n\" (UID: \"e97ceac1-f815-4c26-a0eb-34d882101360\") " pod="openshift-marketplace/community-operators-bcz4n" Oct 03 13:04:56 crc kubenswrapper[4578]: I1003 13:04:56.957467 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e97ceac1-f815-4c26-a0eb-34d882101360-utilities\") pod \"community-operators-bcz4n\" (UID: \"e97ceac1-f815-4c26-a0eb-34d882101360\") " pod="openshift-marketplace/community-operators-bcz4n" Oct 03 13:04:56 crc kubenswrapper[4578]: I1003 13:04:56.957506 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5qzp\" (UniqueName: \"kubernetes.io/projected/e97ceac1-f815-4c26-a0eb-34d882101360-kube-api-access-f5qzp\") pod \"community-operators-bcz4n\" (UID: \"e97ceac1-f815-4c26-a0eb-34d882101360\") " pod="openshift-marketplace/community-operators-bcz4n" Oct 03 13:04:57 crc kubenswrapper[4578]: I1003 13:04:57.059003 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e97ceac1-f815-4c26-a0eb-34d882101360-utilities\") pod \"community-operators-bcz4n\" (UID: \"e97ceac1-f815-4c26-a0eb-34d882101360\") " pod="openshift-marketplace/community-operators-bcz4n" Oct 03 13:04:57 crc kubenswrapper[4578]: I1003 13:04:57.059991 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5qzp\" (UniqueName: \"kubernetes.io/projected/e97ceac1-f815-4c26-a0eb-34d882101360-kube-api-access-f5qzp\") pod \"community-operators-bcz4n\" (UID: \"e97ceac1-f815-4c26-a0eb-34d882101360\") " pod="openshift-marketplace/community-operators-bcz4n" Oct 03 13:04:57 crc kubenswrapper[4578]: I1003 13:04:57.059912 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e97ceac1-f815-4c26-a0eb-34d882101360-utilities\") pod \"community-operators-bcz4n\" (UID: \"e97ceac1-f815-4c26-a0eb-34d882101360\") " pod="openshift-marketplace/community-operators-bcz4n" Oct 03 13:04:57 crc kubenswrapper[4578]: I1003 13:04:57.060109 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e97ceac1-f815-4c26-a0eb-34d882101360-catalog-content\") pod \"community-operators-bcz4n\" (UID: \"e97ceac1-f815-4c26-a0eb-34d882101360\") " pod="openshift-marketplace/community-operators-bcz4n" Oct 03 13:04:57 crc kubenswrapper[4578]: I1003 13:04:57.060611 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e97ceac1-f815-4c26-a0eb-34d882101360-catalog-content\") pod \"community-operators-bcz4n\" (UID: \"e97ceac1-f815-4c26-a0eb-34d882101360\") " pod="openshift-marketplace/community-operators-bcz4n" Oct 03 13:04:57 crc kubenswrapper[4578]: I1003 13:04:57.079300 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5qzp\" (UniqueName: \"kubernetes.io/projected/e97ceac1-f815-4c26-a0eb-34d882101360-kube-api-access-f5qzp\") pod \"community-operators-bcz4n\" (UID: \"e97ceac1-f815-4c26-a0eb-34d882101360\") " pod="openshift-marketplace/community-operators-bcz4n" Oct 03 13:04:57 crc kubenswrapper[4578]: I1003 13:04:57.257613 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bcz4n" Oct 03 13:04:57 crc kubenswrapper[4578]: I1003 13:04:57.541876 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bcz4n"] Oct 03 13:04:57 crc kubenswrapper[4578]: I1003 13:04:57.796855 4578 generic.go:334] "Generic (PLEG): container finished" podID="e97ceac1-f815-4c26-a0eb-34d882101360" containerID="46cc2edb6c1ab2e2f78fc1a4b8d25ec0748fa08b3d3f649949816294289c1a85" exitCode=0 Oct 03 13:04:57 crc kubenswrapper[4578]: I1003 13:04:57.796901 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bcz4n" event={"ID":"e97ceac1-f815-4c26-a0eb-34d882101360","Type":"ContainerDied","Data":"46cc2edb6c1ab2e2f78fc1a4b8d25ec0748fa08b3d3f649949816294289c1a85"} Oct 03 13:04:57 crc kubenswrapper[4578]: I1003 13:04:57.796933 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bcz4n" event={"ID":"e97ceac1-f815-4c26-a0eb-34d882101360","Type":"ContainerStarted","Data":"e80bc0930da41197e3b34ec53d0fe5f4c706fb138f2d4d837bb44320072fd127"} Oct 03 13:04:59 crc kubenswrapper[4578]: I1003 13:04:59.809945 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bcz4n" event={"ID":"e97ceac1-f815-4c26-a0eb-34d882101360","Type":"ContainerStarted","Data":"e40306b7bd2dea1348b611b79ad684c33cef08d89456986c915f4b221f5c6cc7"} Oct 03 13:05:00 crc kubenswrapper[4578]: I1003 13:05:00.816325 4578 generic.go:334] "Generic (PLEG): container finished" podID="e97ceac1-f815-4c26-a0eb-34d882101360" containerID="e40306b7bd2dea1348b611b79ad684c33cef08d89456986c915f4b221f5c6cc7" exitCode=0 Oct 03 13:05:00 crc kubenswrapper[4578]: I1003 13:05:00.816717 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bcz4n" event={"ID":"e97ceac1-f815-4c26-a0eb-34d882101360","Type":"ContainerDied","Data":"e40306b7bd2dea1348b611b79ad684c33cef08d89456986c915f4b221f5c6cc7"} Oct 03 13:05:01 crc kubenswrapper[4578]: I1003 13:05:01.823109 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bcz4n" event={"ID":"e97ceac1-f815-4c26-a0eb-34d882101360","Type":"ContainerStarted","Data":"8d12745b6005a4808b9b6d7b81c74241f620b5dc08d1a385724612258aa19ea1"} Oct 03 13:05:01 crc kubenswrapper[4578]: I1003 13:05:01.840784 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bcz4n" podStartSLOduration=2.40314132 podStartE2EDuration="5.840764477s" podCreationTimestamp="2025-10-03 13:04:56 +0000 UTC" firstStartedPulling="2025-10-03 13:04:57.798227312 +0000 UTC m=+833.596699486" lastFinishedPulling="2025-10-03 13:05:01.235850459 +0000 UTC m=+837.034322643" observedRunningTime="2025-10-03 13:05:01.839466677 +0000 UTC m=+837.637938861" watchObservedRunningTime="2025-10-03 13:05:01.840764477 +0000 UTC m=+837.639236651" Oct 03 13:05:07 crc kubenswrapper[4578]: I1003 13:05:07.258125 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bcz4n" Oct 03 13:05:07 crc kubenswrapper[4578]: I1003 13:05:07.258757 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bcz4n" Oct 03 13:05:07 crc kubenswrapper[4578]: I1003 13:05:07.297541 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bcz4n" Oct 03 13:05:07 crc kubenswrapper[4578]: I1003 13:05:07.331461 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tnrc8"] Oct 03 13:05:07 crc kubenswrapper[4578]: I1003 13:05:07.332844 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tnrc8" Oct 03 13:05:07 crc kubenswrapper[4578]: I1003 13:05:07.338068 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tnrc8"] Oct 03 13:05:07 crc kubenswrapper[4578]: I1003 13:05:07.376052 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mdcf\" (UniqueName: \"kubernetes.io/projected/70804fe7-af6b-4b8e-9731-ce7fd0f5544b-kube-api-access-7mdcf\") pod \"certified-operators-tnrc8\" (UID: \"70804fe7-af6b-4b8e-9731-ce7fd0f5544b\") " pod="openshift-marketplace/certified-operators-tnrc8" Oct 03 13:05:07 crc kubenswrapper[4578]: I1003 13:05:07.376109 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70804fe7-af6b-4b8e-9731-ce7fd0f5544b-utilities\") pod \"certified-operators-tnrc8\" (UID: \"70804fe7-af6b-4b8e-9731-ce7fd0f5544b\") " pod="openshift-marketplace/certified-operators-tnrc8" Oct 03 13:05:07 crc kubenswrapper[4578]: I1003 13:05:07.376129 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70804fe7-af6b-4b8e-9731-ce7fd0f5544b-catalog-content\") pod \"certified-operators-tnrc8\" (UID: \"70804fe7-af6b-4b8e-9731-ce7fd0f5544b\") " pod="openshift-marketplace/certified-operators-tnrc8" Oct 03 13:05:07 crc kubenswrapper[4578]: I1003 13:05:07.476985 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mdcf\" (UniqueName: \"kubernetes.io/projected/70804fe7-af6b-4b8e-9731-ce7fd0f5544b-kube-api-access-7mdcf\") pod \"certified-operators-tnrc8\" (UID: \"70804fe7-af6b-4b8e-9731-ce7fd0f5544b\") " pod="openshift-marketplace/certified-operators-tnrc8" Oct 03 13:05:07 crc kubenswrapper[4578]: I1003 13:05:07.477280 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70804fe7-af6b-4b8e-9731-ce7fd0f5544b-utilities\") pod \"certified-operators-tnrc8\" (UID: \"70804fe7-af6b-4b8e-9731-ce7fd0f5544b\") " pod="openshift-marketplace/certified-operators-tnrc8" Oct 03 13:05:07 crc kubenswrapper[4578]: I1003 13:05:07.477363 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70804fe7-af6b-4b8e-9731-ce7fd0f5544b-catalog-content\") pod \"certified-operators-tnrc8\" (UID: \"70804fe7-af6b-4b8e-9731-ce7fd0f5544b\") " pod="openshift-marketplace/certified-operators-tnrc8" Oct 03 13:05:07 crc kubenswrapper[4578]: I1003 13:05:07.477701 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70804fe7-af6b-4b8e-9731-ce7fd0f5544b-utilities\") pod \"certified-operators-tnrc8\" (UID: \"70804fe7-af6b-4b8e-9731-ce7fd0f5544b\") " pod="openshift-marketplace/certified-operators-tnrc8" Oct 03 13:05:07 crc kubenswrapper[4578]: I1003 13:05:07.477780 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70804fe7-af6b-4b8e-9731-ce7fd0f5544b-catalog-content\") pod \"certified-operators-tnrc8\" (UID: \"70804fe7-af6b-4b8e-9731-ce7fd0f5544b\") " pod="openshift-marketplace/certified-operators-tnrc8" Oct 03 13:05:07 crc kubenswrapper[4578]: I1003 13:05:07.498750 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mdcf\" (UniqueName: \"kubernetes.io/projected/70804fe7-af6b-4b8e-9731-ce7fd0f5544b-kube-api-access-7mdcf\") pod \"certified-operators-tnrc8\" (UID: \"70804fe7-af6b-4b8e-9731-ce7fd0f5544b\") " pod="openshift-marketplace/certified-operators-tnrc8" Oct 03 13:05:07 crc kubenswrapper[4578]: I1003 13:05:07.654484 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tnrc8" Oct 03 13:05:07 crc kubenswrapper[4578]: I1003 13:05:07.922780 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bcz4n" Oct 03 13:05:07 crc kubenswrapper[4578]: I1003 13:05:07.960466 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tnrc8"] Oct 03 13:05:08 crc kubenswrapper[4578]: I1003 13:05:08.866244 4578 generic.go:334] "Generic (PLEG): container finished" podID="70804fe7-af6b-4b8e-9731-ce7fd0f5544b" containerID="df11559491e63fc3dd82a5a1cad14bb6b83f3a6788223f2562b905ba11a20a0f" exitCode=0 Oct 03 13:05:08 crc kubenswrapper[4578]: I1003 13:05:08.866315 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tnrc8" event={"ID":"70804fe7-af6b-4b8e-9731-ce7fd0f5544b","Type":"ContainerDied","Data":"df11559491e63fc3dd82a5a1cad14bb6b83f3a6788223f2562b905ba11a20a0f"} Oct 03 13:05:08 crc kubenswrapper[4578]: I1003 13:05:08.866725 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tnrc8" event={"ID":"70804fe7-af6b-4b8e-9731-ce7fd0f5544b","Type":"ContainerStarted","Data":"7c340277b6cfdb61b0440ba45e3ed17e56c680ad958163ce8cae96c1df1b460e"} Oct 03 13:05:08 crc kubenswrapper[4578]: I1003 13:05:08.897084 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bcz4n"] Oct 03 13:05:09 crc kubenswrapper[4578]: I1003 13:05:09.871290 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bcz4n" podUID="e97ceac1-f815-4c26-a0eb-34d882101360" containerName="registry-server" containerID="cri-o://8d12745b6005a4808b9b6d7b81c74241f620b5dc08d1a385724612258aa19ea1" gracePeriod=2 Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.292501 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bcz4n" Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.410386 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5qzp\" (UniqueName: \"kubernetes.io/projected/e97ceac1-f815-4c26-a0eb-34d882101360-kube-api-access-f5qzp\") pod \"e97ceac1-f815-4c26-a0eb-34d882101360\" (UID: \"e97ceac1-f815-4c26-a0eb-34d882101360\") " Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.410498 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e97ceac1-f815-4c26-a0eb-34d882101360-catalog-content\") pod \"e97ceac1-f815-4c26-a0eb-34d882101360\" (UID: \"e97ceac1-f815-4c26-a0eb-34d882101360\") " Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.410529 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e97ceac1-f815-4c26-a0eb-34d882101360-utilities\") pod \"e97ceac1-f815-4c26-a0eb-34d882101360\" (UID: \"e97ceac1-f815-4c26-a0eb-34d882101360\") " Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.411427 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e97ceac1-f815-4c26-a0eb-34d882101360-utilities" (OuterVolumeSpecName: "utilities") pod "e97ceac1-f815-4c26-a0eb-34d882101360" (UID: "e97ceac1-f815-4c26-a0eb-34d882101360"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.416945 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e97ceac1-f815-4c26-a0eb-34d882101360-kube-api-access-f5qzp" (OuterVolumeSpecName: "kube-api-access-f5qzp") pod "e97ceac1-f815-4c26-a0eb-34d882101360" (UID: "e97ceac1-f815-4c26-a0eb-34d882101360"). InnerVolumeSpecName "kube-api-access-f5qzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.456783 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e97ceac1-f815-4c26-a0eb-34d882101360-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e97ceac1-f815-4c26-a0eb-34d882101360" (UID: "e97ceac1-f815-4c26-a0eb-34d882101360"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.512069 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5qzp\" (UniqueName: \"kubernetes.io/projected/e97ceac1-f815-4c26-a0eb-34d882101360-kube-api-access-f5qzp\") on node \"crc\" DevicePath \"\"" Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.512119 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e97ceac1-f815-4c26-a0eb-34d882101360-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.512130 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e97ceac1-f815-4c26-a0eb-34d882101360-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.879709 4578 generic.go:334] "Generic (PLEG): container finished" podID="e97ceac1-f815-4c26-a0eb-34d882101360" containerID="8d12745b6005a4808b9b6d7b81c74241f620b5dc08d1a385724612258aa19ea1" exitCode=0 Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.879757 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bcz4n" event={"ID":"e97ceac1-f815-4c26-a0eb-34d882101360","Type":"ContainerDied","Data":"8d12745b6005a4808b9b6d7b81c74241f620b5dc08d1a385724612258aa19ea1"} Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.879787 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bcz4n" event={"ID":"e97ceac1-f815-4c26-a0eb-34d882101360","Type":"ContainerDied","Data":"e80bc0930da41197e3b34ec53d0fe5f4c706fb138f2d4d837bb44320072fd127"} Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.879807 4578 scope.go:117] "RemoveContainer" containerID="8d12745b6005a4808b9b6d7b81c74241f620b5dc08d1a385724612258aa19ea1" Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.879933 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bcz4n" Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.918744 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bcz4n"] Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.918779 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bcz4n"] Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.918953 4578 scope.go:117] "RemoveContainer" containerID="e40306b7bd2dea1348b611b79ad684c33cef08d89456986c915f4b221f5c6cc7" Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.945294 4578 scope.go:117] "RemoveContainer" containerID="46cc2edb6c1ab2e2f78fc1a4b8d25ec0748fa08b3d3f649949816294289c1a85" Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.963390 4578 scope.go:117] "RemoveContainer" containerID="8d12745b6005a4808b9b6d7b81c74241f620b5dc08d1a385724612258aa19ea1" Oct 03 13:05:10 crc kubenswrapper[4578]: E1003 13:05:10.963782 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d12745b6005a4808b9b6d7b81c74241f620b5dc08d1a385724612258aa19ea1\": container with ID starting with 8d12745b6005a4808b9b6d7b81c74241f620b5dc08d1a385724612258aa19ea1 not found: ID does not exist" containerID="8d12745b6005a4808b9b6d7b81c74241f620b5dc08d1a385724612258aa19ea1" Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.963826 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d12745b6005a4808b9b6d7b81c74241f620b5dc08d1a385724612258aa19ea1"} err="failed to get container status \"8d12745b6005a4808b9b6d7b81c74241f620b5dc08d1a385724612258aa19ea1\": rpc error: code = NotFound desc = could not find container \"8d12745b6005a4808b9b6d7b81c74241f620b5dc08d1a385724612258aa19ea1\": container with ID starting with 8d12745b6005a4808b9b6d7b81c74241f620b5dc08d1a385724612258aa19ea1 not found: ID does not exist" Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.963852 4578 scope.go:117] "RemoveContainer" containerID="e40306b7bd2dea1348b611b79ad684c33cef08d89456986c915f4b221f5c6cc7" Oct 03 13:05:10 crc kubenswrapper[4578]: E1003 13:05:10.964133 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e40306b7bd2dea1348b611b79ad684c33cef08d89456986c915f4b221f5c6cc7\": container with ID starting with e40306b7bd2dea1348b611b79ad684c33cef08d89456986c915f4b221f5c6cc7 not found: ID does not exist" containerID="e40306b7bd2dea1348b611b79ad684c33cef08d89456986c915f4b221f5c6cc7" Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.964185 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e40306b7bd2dea1348b611b79ad684c33cef08d89456986c915f4b221f5c6cc7"} err="failed to get container status \"e40306b7bd2dea1348b611b79ad684c33cef08d89456986c915f4b221f5c6cc7\": rpc error: code = NotFound desc = could not find container \"e40306b7bd2dea1348b611b79ad684c33cef08d89456986c915f4b221f5c6cc7\": container with ID starting with e40306b7bd2dea1348b611b79ad684c33cef08d89456986c915f4b221f5c6cc7 not found: ID does not exist" Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.964207 4578 scope.go:117] "RemoveContainer" containerID="46cc2edb6c1ab2e2f78fc1a4b8d25ec0748fa08b3d3f649949816294289c1a85" Oct 03 13:05:10 crc kubenswrapper[4578]: E1003 13:05:10.964541 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46cc2edb6c1ab2e2f78fc1a4b8d25ec0748fa08b3d3f649949816294289c1a85\": container with ID starting with 46cc2edb6c1ab2e2f78fc1a4b8d25ec0748fa08b3d3f649949816294289c1a85 not found: ID does not exist" containerID="46cc2edb6c1ab2e2f78fc1a4b8d25ec0748fa08b3d3f649949816294289c1a85" Oct 03 13:05:10 crc kubenswrapper[4578]: I1003 13:05:10.964675 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46cc2edb6c1ab2e2f78fc1a4b8d25ec0748fa08b3d3f649949816294289c1a85"} err="failed to get container status \"46cc2edb6c1ab2e2f78fc1a4b8d25ec0748fa08b3d3f649949816294289c1a85\": rpc error: code = NotFound desc = could not find container \"46cc2edb6c1ab2e2f78fc1a4b8d25ec0748fa08b3d3f649949816294289c1a85\": container with ID starting with 46cc2edb6c1ab2e2f78fc1a4b8d25ec0748fa08b3d3f649949816294289c1a85 not found: ID does not exist" Oct 03 13:05:11 crc kubenswrapper[4578]: I1003 13:05:11.340675 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx"] Oct 03 13:05:11 crc kubenswrapper[4578]: E1003 13:05:11.340865 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e97ceac1-f815-4c26-a0eb-34d882101360" containerName="registry-server" Oct 03 13:05:11 crc kubenswrapper[4578]: I1003 13:05:11.340876 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="e97ceac1-f815-4c26-a0eb-34d882101360" containerName="registry-server" Oct 03 13:05:11 crc kubenswrapper[4578]: E1003 13:05:11.340887 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e97ceac1-f815-4c26-a0eb-34d882101360" containerName="extract-utilities" Oct 03 13:05:11 crc kubenswrapper[4578]: I1003 13:05:11.340892 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="e97ceac1-f815-4c26-a0eb-34d882101360" containerName="extract-utilities" Oct 03 13:05:11 crc kubenswrapper[4578]: E1003 13:05:11.340907 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e97ceac1-f815-4c26-a0eb-34d882101360" containerName="extract-content" Oct 03 13:05:11 crc kubenswrapper[4578]: I1003 13:05:11.340912 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="e97ceac1-f815-4c26-a0eb-34d882101360" containerName="extract-content" Oct 03 13:05:11 crc kubenswrapper[4578]: I1003 13:05:11.341001 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="e97ceac1-f815-4c26-a0eb-34d882101360" containerName="registry-server" Oct 03 13:05:11 crc kubenswrapper[4578]: I1003 13:05:11.341717 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx" Oct 03 13:05:11 crc kubenswrapper[4578]: I1003 13:05:11.346448 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx"] Oct 03 13:05:11 crc kubenswrapper[4578]: I1003 13:05:11.347344 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 03 13:05:11 crc kubenswrapper[4578]: I1003 13:05:11.524782 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw5ct\" (UniqueName: \"kubernetes.io/projected/d75a489d-f32d-4e7b-8f7a-6da21a66340b-kube-api-access-dw5ct\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx\" (UID: \"d75a489d-f32d-4e7b-8f7a-6da21a66340b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx" Oct 03 13:05:11 crc kubenswrapper[4578]: I1003 13:05:11.525079 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d75a489d-f32d-4e7b-8f7a-6da21a66340b-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx\" (UID: \"d75a489d-f32d-4e7b-8f7a-6da21a66340b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx" Oct 03 13:05:11 crc kubenswrapper[4578]: I1003 13:05:11.525130 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d75a489d-f32d-4e7b-8f7a-6da21a66340b-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx\" (UID: \"d75a489d-f32d-4e7b-8f7a-6da21a66340b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx" Oct 03 13:05:11 crc kubenswrapper[4578]: I1003 13:05:11.626038 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d75a489d-f32d-4e7b-8f7a-6da21a66340b-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx\" (UID: \"d75a489d-f32d-4e7b-8f7a-6da21a66340b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx" Oct 03 13:05:11 crc kubenswrapper[4578]: I1003 13:05:11.626113 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d75a489d-f32d-4e7b-8f7a-6da21a66340b-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx\" (UID: \"d75a489d-f32d-4e7b-8f7a-6da21a66340b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx" Oct 03 13:05:11 crc kubenswrapper[4578]: I1003 13:05:11.626571 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d75a489d-f32d-4e7b-8f7a-6da21a66340b-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx\" (UID: \"d75a489d-f32d-4e7b-8f7a-6da21a66340b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx" Oct 03 13:05:11 crc kubenswrapper[4578]: I1003 13:05:11.626653 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d75a489d-f32d-4e7b-8f7a-6da21a66340b-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx\" (UID: \"d75a489d-f32d-4e7b-8f7a-6da21a66340b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx" Oct 03 13:05:11 crc kubenswrapper[4578]: I1003 13:05:11.626783 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw5ct\" (UniqueName: \"kubernetes.io/projected/d75a489d-f32d-4e7b-8f7a-6da21a66340b-kube-api-access-dw5ct\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx\" (UID: \"d75a489d-f32d-4e7b-8f7a-6da21a66340b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx" Oct 03 13:05:11 crc kubenswrapper[4578]: I1003 13:05:11.657611 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw5ct\" (UniqueName: \"kubernetes.io/projected/d75a489d-f32d-4e7b-8f7a-6da21a66340b-kube-api-access-dw5ct\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx\" (UID: \"d75a489d-f32d-4e7b-8f7a-6da21a66340b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx" Oct 03 13:05:11 crc kubenswrapper[4578]: I1003 13:05:11.660819 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx" Oct 03 13:05:11 crc kubenswrapper[4578]: I1003 13:05:11.901089 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx"] Oct 03 13:05:12 crc kubenswrapper[4578]: I1003 13:05:12.893586 4578 generic.go:334] "Generic (PLEG): container finished" podID="d75a489d-f32d-4e7b-8f7a-6da21a66340b" containerID="f98d031e04b9d2b0adcc7112dbef5816a543cbfb2ed899f382e3f8f091236ea4" exitCode=0 Oct 03 13:05:12 crc kubenswrapper[4578]: I1003 13:05:12.893702 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx" event={"ID":"d75a489d-f32d-4e7b-8f7a-6da21a66340b","Type":"ContainerDied","Data":"f98d031e04b9d2b0adcc7112dbef5816a543cbfb2ed899f382e3f8f091236ea4"} Oct 03 13:05:12 crc kubenswrapper[4578]: I1003 13:05:12.893920 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx" event={"ID":"d75a489d-f32d-4e7b-8f7a-6da21a66340b","Type":"ContainerStarted","Data":"0549d4687ced8ad982e257effd335e60f1f7476cacccba3ad120a82952cb08b0"} Oct 03 13:05:12 crc kubenswrapper[4578]: I1003 13:05:12.918392 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e97ceac1-f815-4c26-a0eb-34d882101360" path="/var/lib/kubelet/pods/e97ceac1-f815-4c26-a0eb-34d882101360/volumes" Oct 03 13:05:13 crc kubenswrapper[4578]: I1003 13:05:13.901453 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tnrc8" event={"ID":"70804fe7-af6b-4b8e-9731-ce7fd0f5544b","Type":"ContainerStarted","Data":"1deec3d7ce85f8472d19975847c26523c797fe5c682f214abcb6b951017e3028"} Oct 03 13:05:14 crc kubenswrapper[4578]: I1003 13:05:14.309595 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-b4g4t"] Oct 03 13:05:14 crc kubenswrapper[4578]: I1003 13:05:14.320060 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b4g4t"] Oct 03 13:05:14 crc kubenswrapper[4578]: I1003 13:05:14.320216 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b4g4t" Oct 03 13:05:14 crc kubenswrapper[4578]: I1003 13:05:14.463359 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c92958b-5203-4955-b58a-0869cd4560a5-catalog-content\") pod \"redhat-operators-b4g4t\" (UID: \"8c92958b-5203-4955-b58a-0869cd4560a5\") " pod="openshift-marketplace/redhat-operators-b4g4t" Oct 03 13:05:14 crc kubenswrapper[4578]: I1003 13:05:14.463411 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzm6r\" (UniqueName: \"kubernetes.io/projected/8c92958b-5203-4955-b58a-0869cd4560a5-kube-api-access-pzm6r\") pod \"redhat-operators-b4g4t\" (UID: \"8c92958b-5203-4955-b58a-0869cd4560a5\") " pod="openshift-marketplace/redhat-operators-b4g4t" Oct 03 13:05:14 crc kubenswrapper[4578]: I1003 13:05:14.463493 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c92958b-5203-4955-b58a-0869cd4560a5-utilities\") pod \"redhat-operators-b4g4t\" (UID: \"8c92958b-5203-4955-b58a-0869cd4560a5\") " pod="openshift-marketplace/redhat-operators-b4g4t" Oct 03 13:05:14 crc kubenswrapper[4578]: I1003 13:05:14.565062 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c92958b-5203-4955-b58a-0869cd4560a5-catalog-content\") pod \"redhat-operators-b4g4t\" (UID: \"8c92958b-5203-4955-b58a-0869cd4560a5\") " pod="openshift-marketplace/redhat-operators-b4g4t" Oct 03 13:05:14 crc kubenswrapper[4578]: I1003 13:05:14.565100 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzm6r\" (UniqueName: \"kubernetes.io/projected/8c92958b-5203-4955-b58a-0869cd4560a5-kube-api-access-pzm6r\") pod \"redhat-operators-b4g4t\" (UID: \"8c92958b-5203-4955-b58a-0869cd4560a5\") " pod="openshift-marketplace/redhat-operators-b4g4t" Oct 03 13:05:14 crc kubenswrapper[4578]: I1003 13:05:14.565162 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c92958b-5203-4955-b58a-0869cd4560a5-utilities\") pod \"redhat-operators-b4g4t\" (UID: \"8c92958b-5203-4955-b58a-0869cd4560a5\") " pod="openshift-marketplace/redhat-operators-b4g4t" Oct 03 13:05:14 crc kubenswrapper[4578]: I1003 13:05:14.565647 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c92958b-5203-4955-b58a-0869cd4560a5-catalog-content\") pod \"redhat-operators-b4g4t\" (UID: \"8c92958b-5203-4955-b58a-0869cd4560a5\") " pod="openshift-marketplace/redhat-operators-b4g4t" Oct 03 13:05:14 crc kubenswrapper[4578]: I1003 13:05:14.565698 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c92958b-5203-4955-b58a-0869cd4560a5-utilities\") pod \"redhat-operators-b4g4t\" (UID: \"8c92958b-5203-4955-b58a-0869cd4560a5\") " pod="openshift-marketplace/redhat-operators-b4g4t" Oct 03 13:05:14 crc kubenswrapper[4578]: I1003 13:05:14.584038 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzm6r\" (UniqueName: \"kubernetes.io/projected/8c92958b-5203-4955-b58a-0869cd4560a5-kube-api-access-pzm6r\") pod \"redhat-operators-b4g4t\" (UID: \"8c92958b-5203-4955-b58a-0869cd4560a5\") " pod="openshift-marketplace/redhat-operators-b4g4t" Oct 03 13:05:14 crc kubenswrapper[4578]: I1003 13:05:14.670134 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b4g4t" Oct 03 13:05:14 crc kubenswrapper[4578]: I1003 13:05:14.888806 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b4g4t"] Oct 03 13:05:14 crc kubenswrapper[4578]: I1003 13:05:14.917325 4578 generic.go:334] "Generic (PLEG): container finished" podID="70804fe7-af6b-4b8e-9731-ce7fd0f5544b" containerID="1deec3d7ce85f8472d19975847c26523c797fe5c682f214abcb6b951017e3028" exitCode=0 Oct 03 13:05:14 crc kubenswrapper[4578]: I1003 13:05:14.921275 4578 generic.go:334] "Generic (PLEG): container finished" podID="d75a489d-f32d-4e7b-8f7a-6da21a66340b" containerID="8f3947c3df13ab0c66ffeb9e8197eab6d298e2f90bc13d5a0dc6fa0e75713056" exitCode=0 Oct 03 13:05:14 crc kubenswrapper[4578]: I1003 13:05:14.944915 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tnrc8" event={"ID":"70804fe7-af6b-4b8e-9731-ce7fd0f5544b","Type":"ContainerDied","Data":"1deec3d7ce85f8472d19975847c26523c797fe5c682f214abcb6b951017e3028"} Oct 03 13:05:14 crc kubenswrapper[4578]: I1003 13:05:14.944966 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b4g4t" event={"ID":"8c92958b-5203-4955-b58a-0869cd4560a5","Type":"ContainerStarted","Data":"603e768d0128dd051c26180d5adf0cd0f1359210195c871fa19d9b091f5fc702"} Oct 03 13:05:14 crc kubenswrapper[4578]: I1003 13:05:14.944981 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx" event={"ID":"d75a489d-f32d-4e7b-8f7a-6da21a66340b","Type":"ContainerDied","Data":"8f3947c3df13ab0c66ffeb9e8197eab6d298e2f90bc13d5a0dc6fa0e75713056"} Oct 03 13:05:15 crc kubenswrapper[4578]: I1003 13:05:15.928526 4578 generic.go:334] "Generic (PLEG): container finished" podID="d75a489d-f32d-4e7b-8f7a-6da21a66340b" containerID="c8b27f47e376ebc6bf9a33acf56f6a64784adea27f277a17840ac700d695749d" exitCode=0 Oct 03 13:05:15 crc kubenswrapper[4578]: I1003 13:05:15.928582 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx" event={"ID":"d75a489d-f32d-4e7b-8f7a-6da21a66340b","Type":"ContainerDied","Data":"c8b27f47e376ebc6bf9a33acf56f6a64784adea27f277a17840ac700d695749d"} Oct 03 13:05:15 crc kubenswrapper[4578]: I1003 13:05:15.931258 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tnrc8" event={"ID":"70804fe7-af6b-4b8e-9731-ce7fd0f5544b","Type":"ContainerStarted","Data":"c48740ccbfb535ea627e2146a1c60c283ca47aa53e141b57e700c79cdee6948c"} Oct 03 13:05:15 crc kubenswrapper[4578]: I1003 13:05:15.932763 4578 generic.go:334] "Generic (PLEG): container finished" podID="8c92958b-5203-4955-b58a-0869cd4560a5" containerID="19608660184ce9b40b6ef8086157aff4e253f41d68ce6a3b8ce52d98d27785f6" exitCode=0 Oct 03 13:05:15 crc kubenswrapper[4578]: I1003 13:05:15.932790 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b4g4t" event={"ID":"8c92958b-5203-4955-b58a-0869cd4560a5","Type":"ContainerDied","Data":"19608660184ce9b40b6ef8086157aff4e253f41d68ce6a3b8ce52d98d27785f6"} Oct 03 13:05:17 crc kubenswrapper[4578]: I1003 13:05:17.165375 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx" Oct 03 13:05:17 crc kubenswrapper[4578]: I1003 13:05:17.182346 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tnrc8" podStartSLOduration=3.647863809 podStartE2EDuration="10.182319167s" podCreationTimestamp="2025-10-03 13:05:07 +0000 UTC" firstStartedPulling="2025-10-03 13:05:08.86792427 +0000 UTC m=+844.666396454" lastFinishedPulling="2025-10-03 13:05:15.402379628 +0000 UTC m=+851.200851812" observedRunningTime="2025-10-03 13:05:15.987125645 +0000 UTC m=+851.785597829" watchObservedRunningTime="2025-10-03 13:05:17.182319167 +0000 UTC m=+852.980791351" Oct 03 13:05:17 crc kubenswrapper[4578]: I1003 13:05:17.301495 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dw5ct\" (UniqueName: \"kubernetes.io/projected/d75a489d-f32d-4e7b-8f7a-6da21a66340b-kube-api-access-dw5ct\") pod \"d75a489d-f32d-4e7b-8f7a-6da21a66340b\" (UID: \"d75a489d-f32d-4e7b-8f7a-6da21a66340b\") " Oct 03 13:05:17 crc kubenswrapper[4578]: I1003 13:05:17.301573 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d75a489d-f32d-4e7b-8f7a-6da21a66340b-bundle\") pod \"d75a489d-f32d-4e7b-8f7a-6da21a66340b\" (UID: \"d75a489d-f32d-4e7b-8f7a-6da21a66340b\") " Oct 03 13:05:17 crc kubenswrapper[4578]: I1003 13:05:17.301614 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d75a489d-f32d-4e7b-8f7a-6da21a66340b-util\") pod \"d75a489d-f32d-4e7b-8f7a-6da21a66340b\" (UID: \"d75a489d-f32d-4e7b-8f7a-6da21a66340b\") " Oct 03 13:05:17 crc kubenswrapper[4578]: I1003 13:05:17.302193 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d75a489d-f32d-4e7b-8f7a-6da21a66340b-bundle" (OuterVolumeSpecName: "bundle") pod "d75a489d-f32d-4e7b-8f7a-6da21a66340b" (UID: "d75a489d-f32d-4e7b-8f7a-6da21a66340b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:05:17 crc kubenswrapper[4578]: I1003 13:05:17.310271 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d75a489d-f32d-4e7b-8f7a-6da21a66340b-kube-api-access-dw5ct" (OuterVolumeSpecName: "kube-api-access-dw5ct") pod "d75a489d-f32d-4e7b-8f7a-6da21a66340b" (UID: "d75a489d-f32d-4e7b-8f7a-6da21a66340b"). InnerVolumeSpecName "kube-api-access-dw5ct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:05:17 crc kubenswrapper[4578]: I1003 13:05:17.402723 4578 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d75a489d-f32d-4e7b-8f7a-6da21a66340b-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:05:17 crc kubenswrapper[4578]: I1003 13:05:17.402755 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dw5ct\" (UniqueName: \"kubernetes.io/projected/d75a489d-f32d-4e7b-8f7a-6da21a66340b-kube-api-access-dw5ct\") on node \"crc\" DevicePath \"\"" Oct 03 13:05:17 crc kubenswrapper[4578]: I1003 13:05:17.591840 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d75a489d-f32d-4e7b-8f7a-6da21a66340b-util" (OuterVolumeSpecName: "util") pod "d75a489d-f32d-4e7b-8f7a-6da21a66340b" (UID: "d75a489d-f32d-4e7b-8f7a-6da21a66340b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:05:17 crc kubenswrapper[4578]: I1003 13:05:17.606291 4578 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d75a489d-f32d-4e7b-8f7a-6da21a66340b-util\") on node \"crc\" DevicePath \"\"" Oct 03 13:05:17 crc kubenswrapper[4578]: I1003 13:05:17.655469 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tnrc8" Oct 03 13:05:17 crc kubenswrapper[4578]: I1003 13:05:17.655521 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tnrc8" Oct 03 13:05:17 crc kubenswrapper[4578]: I1003 13:05:17.695596 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tnrc8" Oct 03 13:05:17 crc kubenswrapper[4578]: I1003 13:05:17.944221 4578 generic.go:334] "Generic (PLEG): container finished" podID="8c92958b-5203-4955-b58a-0869cd4560a5" containerID="918a8245b3377a4220b0167bd88f26327bbbe72bddac7fd3d2bba7700bb98ef2" exitCode=0 Oct 03 13:05:17 crc kubenswrapper[4578]: I1003 13:05:17.944289 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b4g4t" event={"ID":"8c92958b-5203-4955-b58a-0869cd4560a5","Type":"ContainerDied","Data":"918a8245b3377a4220b0167bd88f26327bbbe72bddac7fd3d2bba7700bb98ef2"} Oct 03 13:05:17 crc kubenswrapper[4578]: I1003 13:05:17.949020 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx" event={"ID":"d75a489d-f32d-4e7b-8f7a-6da21a66340b","Type":"ContainerDied","Data":"0549d4687ced8ad982e257effd335e60f1f7476cacccba3ad120a82952cb08b0"} Oct 03 13:05:17 crc kubenswrapper[4578]: I1003 13:05:17.949052 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0549d4687ced8ad982e257effd335e60f1f7476cacccba3ad120a82952cb08b0" Oct 03 13:05:17 crc kubenswrapper[4578]: I1003 13:05:17.949051 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx" Oct 03 13:05:20 crc kubenswrapper[4578]: I1003 13:05:20.411950 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-54fwn"] Oct 03 13:05:20 crc kubenswrapper[4578]: E1003 13:05:20.412517 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d75a489d-f32d-4e7b-8f7a-6da21a66340b" containerName="util" Oct 03 13:05:20 crc kubenswrapper[4578]: I1003 13:05:20.412532 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="d75a489d-f32d-4e7b-8f7a-6da21a66340b" containerName="util" Oct 03 13:05:20 crc kubenswrapper[4578]: E1003 13:05:20.412547 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d75a489d-f32d-4e7b-8f7a-6da21a66340b" containerName="extract" Oct 03 13:05:20 crc kubenswrapper[4578]: I1003 13:05:20.412556 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="d75a489d-f32d-4e7b-8f7a-6da21a66340b" containerName="extract" Oct 03 13:05:20 crc kubenswrapper[4578]: E1003 13:05:20.412578 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d75a489d-f32d-4e7b-8f7a-6da21a66340b" containerName="pull" Oct 03 13:05:20 crc kubenswrapper[4578]: I1003 13:05:20.412594 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="d75a489d-f32d-4e7b-8f7a-6da21a66340b" containerName="pull" Oct 03 13:05:20 crc kubenswrapper[4578]: I1003 13:05:20.412767 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="d75a489d-f32d-4e7b-8f7a-6da21a66340b" containerName="extract" Oct 03 13:05:20 crc kubenswrapper[4578]: I1003 13:05:20.413217 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-54fwn" Oct 03 13:05:20 crc kubenswrapper[4578]: I1003 13:05:20.414864 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-stw8t" Oct 03 13:05:20 crc kubenswrapper[4578]: I1003 13:05:20.416431 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 03 13:05:20 crc kubenswrapper[4578]: I1003 13:05:20.417526 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 03 13:05:20 crc kubenswrapper[4578]: I1003 13:05:20.428300 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-54fwn"] Oct 03 13:05:20 crc kubenswrapper[4578]: I1003 13:05:20.541358 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrbnx\" (UniqueName: \"kubernetes.io/projected/566708f1-bb6e-4f0b-b381-d594f60b864b-kube-api-access-qrbnx\") pod \"nmstate-operator-858ddd8f98-54fwn\" (UID: \"566708f1-bb6e-4f0b-b381-d594f60b864b\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-54fwn" Oct 03 13:05:20 crc kubenswrapper[4578]: I1003 13:05:20.642658 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrbnx\" (UniqueName: \"kubernetes.io/projected/566708f1-bb6e-4f0b-b381-d594f60b864b-kube-api-access-qrbnx\") pod \"nmstate-operator-858ddd8f98-54fwn\" (UID: \"566708f1-bb6e-4f0b-b381-d594f60b864b\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-54fwn" Oct 03 13:05:20 crc kubenswrapper[4578]: I1003 13:05:20.688653 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrbnx\" (UniqueName: \"kubernetes.io/projected/566708f1-bb6e-4f0b-b381-d594f60b864b-kube-api-access-qrbnx\") pod \"nmstate-operator-858ddd8f98-54fwn\" (UID: \"566708f1-bb6e-4f0b-b381-d594f60b864b\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-54fwn" Oct 03 13:05:20 crc kubenswrapper[4578]: I1003 13:05:20.731242 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-54fwn" Oct 03 13:05:20 crc kubenswrapper[4578]: I1003 13:05:20.964749 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b4g4t" event={"ID":"8c92958b-5203-4955-b58a-0869cd4560a5","Type":"ContainerStarted","Data":"5d10497f11b7ec48957c604b4a7648ad61c90eac6b368cb8f4a49db6cbd873cc"} Oct 03 13:05:20 crc kubenswrapper[4578]: I1003 13:05:20.985387 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-b4g4t" podStartSLOduration=2.734406414 podStartE2EDuration="6.985364504s" podCreationTimestamp="2025-10-03 13:05:14 +0000 UTC" firstStartedPulling="2025-10-03 13:05:15.933819294 +0000 UTC m=+851.732291478" lastFinishedPulling="2025-10-03 13:05:20.184777384 +0000 UTC m=+855.983249568" observedRunningTime="2025-10-03 13:05:20.982227579 +0000 UTC m=+856.780699783" watchObservedRunningTime="2025-10-03 13:05:20.985364504 +0000 UTC m=+856.783836688" Oct 03 13:05:21 crc kubenswrapper[4578]: I1003 13:05:21.152236 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-54fwn"] Oct 03 13:05:21 crc kubenswrapper[4578]: I1003 13:05:21.296770 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zlx4b"] Oct 03 13:05:21 crc kubenswrapper[4578]: I1003 13:05:21.298029 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zlx4b" Oct 03 13:05:21 crc kubenswrapper[4578]: I1003 13:05:21.308813 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zlx4b"] Oct 03 13:05:21 crc kubenswrapper[4578]: I1003 13:05:21.454669 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxn57\" (UniqueName: \"kubernetes.io/projected/15764dfd-829c-45fb-bc6c-7a7e918d6a3c-kube-api-access-kxn57\") pod \"redhat-marketplace-zlx4b\" (UID: \"15764dfd-829c-45fb-bc6c-7a7e918d6a3c\") " pod="openshift-marketplace/redhat-marketplace-zlx4b" Oct 03 13:05:21 crc kubenswrapper[4578]: I1003 13:05:21.454750 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15764dfd-829c-45fb-bc6c-7a7e918d6a3c-catalog-content\") pod \"redhat-marketplace-zlx4b\" (UID: \"15764dfd-829c-45fb-bc6c-7a7e918d6a3c\") " pod="openshift-marketplace/redhat-marketplace-zlx4b" Oct 03 13:05:21 crc kubenswrapper[4578]: I1003 13:05:21.454778 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15764dfd-829c-45fb-bc6c-7a7e918d6a3c-utilities\") pod \"redhat-marketplace-zlx4b\" (UID: \"15764dfd-829c-45fb-bc6c-7a7e918d6a3c\") " pod="openshift-marketplace/redhat-marketplace-zlx4b" Oct 03 13:05:21 crc kubenswrapper[4578]: I1003 13:05:21.556061 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxn57\" (UniqueName: \"kubernetes.io/projected/15764dfd-829c-45fb-bc6c-7a7e918d6a3c-kube-api-access-kxn57\") pod \"redhat-marketplace-zlx4b\" (UID: \"15764dfd-829c-45fb-bc6c-7a7e918d6a3c\") " pod="openshift-marketplace/redhat-marketplace-zlx4b" Oct 03 13:05:21 crc kubenswrapper[4578]: I1003 13:05:21.556139 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15764dfd-829c-45fb-bc6c-7a7e918d6a3c-catalog-content\") pod \"redhat-marketplace-zlx4b\" (UID: \"15764dfd-829c-45fb-bc6c-7a7e918d6a3c\") " pod="openshift-marketplace/redhat-marketplace-zlx4b" Oct 03 13:05:21 crc kubenswrapper[4578]: I1003 13:05:21.556163 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15764dfd-829c-45fb-bc6c-7a7e918d6a3c-utilities\") pod \"redhat-marketplace-zlx4b\" (UID: \"15764dfd-829c-45fb-bc6c-7a7e918d6a3c\") " pod="openshift-marketplace/redhat-marketplace-zlx4b" Oct 03 13:05:21 crc kubenswrapper[4578]: I1003 13:05:21.556661 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15764dfd-829c-45fb-bc6c-7a7e918d6a3c-utilities\") pod \"redhat-marketplace-zlx4b\" (UID: \"15764dfd-829c-45fb-bc6c-7a7e918d6a3c\") " pod="openshift-marketplace/redhat-marketplace-zlx4b" Oct 03 13:05:21 crc kubenswrapper[4578]: I1003 13:05:21.556681 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15764dfd-829c-45fb-bc6c-7a7e918d6a3c-catalog-content\") pod \"redhat-marketplace-zlx4b\" (UID: \"15764dfd-829c-45fb-bc6c-7a7e918d6a3c\") " pod="openshift-marketplace/redhat-marketplace-zlx4b" Oct 03 13:05:21 crc kubenswrapper[4578]: I1003 13:05:21.573970 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxn57\" (UniqueName: \"kubernetes.io/projected/15764dfd-829c-45fb-bc6c-7a7e918d6a3c-kube-api-access-kxn57\") pod \"redhat-marketplace-zlx4b\" (UID: \"15764dfd-829c-45fb-bc6c-7a7e918d6a3c\") " pod="openshift-marketplace/redhat-marketplace-zlx4b" Oct 03 13:05:21 crc kubenswrapper[4578]: I1003 13:05:21.612512 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zlx4b" Oct 03 13:05:21 crc kubenswrapper[4578]: I1003 13:05:21.914485 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zlx4b"] Oct 03 13:05:21 crc kubenswrapper[4578]: W1003 13:05:21.919520 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15764dfd_829c_45fb_bc6c_7a7e918d6a3c.slice/crio-623b28423b49f2b7b9ef5ec2e4f2f96980cbf8ba8fbe453a6ff635c3bb95fed6 WatchSource:0}: Error finding container 623b28423b49f2b7b9ef5ec2e4f2f96980cbf8ba8fbe453a6ff635c3bb95fed6: Status 404 returned error can't find the container with id 623b28423b49f2b7b9ef5ec2e4f2f96980cbf8ba8fbe453a6ff635c3bb95fed6 Oct 03 13:05:21 crc kubenswrapper[4578]: I1003 13:05:21.970625 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-54fwn" event={"ID":"566708f1-bb6e-4f0b-b381-d594f60b864b","Type":"ContainerStarted","Data":"db31abca67ad0455cdd648e32e71a346307359a1cc96e8a40e931c582f475163"} Oct 03 13:05:21 crc kubenswrapper[4578]: I1003 13:05:21.971774 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zlx4b" event={"ID":"15764dfd-829c-45fb-bc6c-7a7e918d6a3c","Type":"ContainerStarted","Data":"623b28423b49f2b7b9ef5ec2e4f2f96980cbf8ba8fbe453a6ff635c3bb95fed6"} Oct 03 13:05:22 crc kubenswrapper[4578]: I1003 13:05:22.979099 4578 generic.go:334] "Generic (PLEG): container finished" podID="15764dfd-829c-45fb-bc6c-7a7e918d6a3c" containerID="26787cdd8a79598548808e3c514cdd26cc69104804931dab12c2c02e0c7cddd5" exitCode=0 Oct 03 13:05:22 crc kubenswrapper[4578]: I1003 13:05:22.979147 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zlx4b" event={"ID":"15764dfd-829c-45fb-bc6c-7a7e918d6a3c","Type":"ContainerDied","Data":"26787cdd8a79598548808e3c514cdd26cc69104804931dab12c2c02e0c7cddd5"} Oct 03 13:05:24 crc kubenswrapper[4578]: I1003 13:05:24.670346 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-b4g4t" Oct 03 13:05:24 crc kubenswrapper[4578]: I1003 13:05:24.670929 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-b4g4t" Oct 03 13:05:24 crc kubenswrapper[4578]: I1003 13:05:24.990523 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-54fwn" event={"ID":"566708f1-bb6e-4f0b-b381-d594f60b864b","Type":"ContainerStarted","Data":"567989248b76edeb7fe22be75472cd21b0907601b573230944a6518101cdf91c"} Oct 03 13:05:24 crc kubenswrapper[4578]: I1003 13:05:24.992664 4578 generic.go:334] "Generic (PLEG): container finished" podID="15764dfd-829c-45fb-bc6c-7a7e918d6a3c" containerID="4d9960ead89232c705a3d0e8cb83e4e941fd1e5c72f973c02ffb9350369782b8" exitCode=0 Oct 03 13:05:24 crc kubenswrapper[4578]: I1003 13:05:24.992720 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zlx4b" event={"ID":"15764dfd-829c-45fb-bc6c-7a7e918d6a3c","Type":"ContainerDied","Data":"4d9960ead89232c705a3d0e8cb83e4e941fd1e5c72f973c02ffb9350369782b8"} Oct 03 13:05:25 crc kubenswrapper[4578]: I1003 13:05:25.006283 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-54fwn" podStartSLOduration=2.414500394 podStartE2EDuration="5.006268133s" podCreationTimestamp="2025-10-03 13:05:20 +0000 UTC" firstStartedPulling="2025-10-03 13:05:21.154123342 +0000 UTC m=+856.952595526" lastFinishedPulling="2025-10-03 13:05:23.745891081 +0000 UTC m=+859.544363265" observedRunningTime="2025-10-03 13:05:25.003339295 +0000 UTC m=+860.801811479" watchObservedRunningTime="2025-10-03 13:05:25.006268133 +0000 UTC m=+860.804740317" Oct 03 13:05:25 crc kubenswrapper[4578]: I1003 13:05:25.706354 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-b4g4t" podUID="8c92958b-5203-4955-b58a-0869cd4560a5" containerName="registry-server" probeResult="failure" output=< Oct 03 13:05:25 crc kubenswrapper[4578]: timeout: failed to connect service ":50051" within 1s Oct 03 13:05:25 crc kubenswrapper[4578]: > Oct 03 13:05:26 crc kubenswrapper[4578]: I1003 13:05:26.000957 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zlx4b" event={"ID":"15764dfd-829c-45fb-bc6c-7a7e918d6a3c","Type":"ContainerStarted","Data":"009ff34b913cdedd4457f95fa360305de69f925ff6d45d342f1c00f3618780eb"} Oct 03 13:05:26 crc kubenswrapper[4578]: I1003 13:05:26.022449 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zlx4b" podStartSLOduration=2.606138269 podStartE2EDuration="5.022431417s" podCreationTimestamp="2025-10-03 13:05:21 +0000 UTC" firstStartedPulling="2025-10-03 13:05:22.981028751 +0000 UTC m=+858.779500935" lastFinishedPulling="2025-10-03 13:05:25.397321899 +0000 UTC m=+861.195794083" observedRunningTime="2025-10-03 13:05:26.015999162 +0000 UTC m=+861.814471366" watchObservedRunningTime="2025-10-03 13:05:26.022431417 +0000 UTC m=+861.820903601" Oct 03 13:05:27 crc kubenswrapper[4578]: I1003 13:05:27.693543 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tnrc8" Oct 03 13:05:30 crc kubenswrapper[4578]: I1003 13:05:30.962727 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tnrc8"] Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.222014 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-652tz"] Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.222918 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-652tz" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.225078 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-wz2pz" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.232428 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-fp4rx"] Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.233292 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-fp4rx" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.236148 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.241400 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-652tz"] Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.255946 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-vknp5"] Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.256619 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-vknp5" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.278101 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-fp4rx"] Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.371099 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b111b867-4d61-489f-aa89-f820b50c9a92-nmstate-lock\") pod \"nmstate-handler-vknp5\" (UID: \"b111b867-4d61-489f-aa89-f820b50c9a92\") " pod="openshift-nmstate/nmstate-handler-vknp5" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.371141 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b111b867-4d61-489f-aa89-f820b50c9a92-ovs-socket\") pod \"nmstate-handler-vknp5\" (UID: \"b111b867-4d61-489f-aa89-f820b50c9a92\") " pod="openshift-nmstate/nmstate-handler-vknp5" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.371168 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b111b867-4d61-489f-aa89-f820b50c9a92-dbus-socket\") pod \"nmstate-handler-vknp5\" (UID: \"b111b867-4d61-489f-aa89-f820b50c9a92\") " pod="openshift-nmstate/nmstate-handler-vknp5" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.371196 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bf9f48d4-44a0-4a43-872d-fce41608a68e-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-fp4rx\" (UID: \"bf9f48d4-44a0-4a43-872d-fce41608a68e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-fp4rx" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.371216 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67bqv\" (UniqueName: \"kubernetes.io/projected/bf9f48d4-44a0-4a43-872d-fce41608a68e-kube-api-access-67bqv\") pod \"nmstate-webhook-6cdbc54649-fp4rx\" (UID: \"bf9f48d4-44a0-4a43-872d-fce41608a68e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-fp4rx" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.371238 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7thx\" (UniqueName: \"kubernetes.io/projected/b111b867-4d61-489f-aa89-f820b50c9a92-kube-api-access-k7thx\") pod \"nmstate-handler-vknp5\" (UID: \"b111b867-4d61-489f-aa89-f820b50c9a92\") " pod="openshift-nmstate/nmstate-handler-vknp5" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.371288 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btqfx\" (UniqueName: \"kubernetes.io/projected/94dd97ab-9637-4eab-82dc-5d070129abac-kube-api-access-btqfx\") pod \"nmstate-metrics-fdff9cb8d-652tz\" (UID: \"94dd97ab-9637-4eab-82dc-5d070129abac\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-652tz" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.385844 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-4xjn2"] Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.393677 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4xjn2" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.396271 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-4xjn2"] Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.396421 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.396744 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-mstdd" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.397423 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.472105 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b111b867-4d61-489f-aa89-f820b50c9a92-dbus-socket\") pod \"nmstate-handler-vknp5\" (UID: \"b111b867-4d61-489f-aa89-f820b50c9a92\") " pod="openshift-nmstate/nmstate-handler-vknp5" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.472159 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bf9f48d4-44a0-4a43-872d-fce41608a68e-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-fp4rx\" (UID: \"bf9f48d4-44a0-4a43-872d-fce41608a68e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-fp4rx" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.472184 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67bqv\" (UniqueName: \"kubernetes.io/projected/bf9f48d4-44a0-4a43-872d-fce41608a68e-kube-api-access-67bqv\") pod \"nmstate-webhook-6cdbc54649-fp4rx\" (UID: \"bf9f48d4-44a0-4a43-872d-fce41608a68e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-fp4rx" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.472203 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7thx\" (UniqueName: \"kubernetes.io/projected/b111b867-4d61-489f-aa89-f820b50c9a92-kube-api-access-k7thx\") pod \"nmstate-handler-vknp5\" (UID: \"b111b867-4d61-489f-aa89-f820b50c9a92\") " pod="openshift-nmstate/nmstate-handler-vknp5" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.472254 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btqfx\" (UniqueName: \"kubernetes.io/projected/94dd97ab-9637-4eab-82dc-5d070129abac-kube-api-access-btqfx\") pod \"nmstate-metrics-fdff9cb8d-652tz\" (UID: \"94dd97ab-9637-4eab-82dc-5d070129abac\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-652tz" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.472274 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b111b867-4d61-489f-aa89-f820b50c9a92-nmstate-lock\") pod \"nmstate-handler-vknp5\" (UID: \"b111b867-4d61-489f-aa89-f820b50c9a92\") " pod="openshift-nmstate/nmstate-handler-vknp5" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.472290 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b111b867-4d61-489f-aa89-f820b50c9a92-ovs-socket\") pod \"nmstate-handler-vknp5\" (UID: \"b111b867-4d61-489f-aa89-f820b50c9a92\") " pod="openshift-nmstate/nmstate-handler-vknp5" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.472349 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b111b867-4d61-489f-aa89-f820b50c9a92-ovs-socket\") pod \"nmstate-handler-vknp5\" (UID: \"b111b867-4d61-489f-aa89-f820b50c9a92\") " pod="openshift-nmstate/nmstate-handler-vknp5" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.472470 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b111b867-4d61-489f-aa89-f820b50c9a92-dbus-socket\") pod \"nmstate-handler-vknp5\" (UID: \"b111b867-4d61-489f-aa89-f820b50c9a92\") " pod="openshift-nmstate/nmstate-handler-vknp5" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.472606 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b111b867-4d61-489f-aa89-f820b50c9a92-nmstate-lock\") pod \"nmstate-handler-vknp5\" (UID: \"b111b867-4d61-489f-aa89-f820b50c9a92\") " pod="openshift-nmstate/nmstate-handler-vknp5" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.480385 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bf9f48d4-44a0-4a43-872d-fce41608a68e-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-fp4rx\" (UID: \"bf9f48d4-44a0-4a43-872d-fce41608a68e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-fp4rx" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.488681 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67bqv\" (UniqueName: \"kubernetes.io/projected/bf9f48d4-44a0-4a43-872d-fce41608a68e-kube-api-access-67bqv\") pod \"nmstate-webhook-6cdbc54649-fp4rx\" (UID: \"bf9f48d4-44a0-4a43-872d-fce41608a68e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-fp4rx" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.501937 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x2b49"] Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.502143 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-x2b49" podUID="d5afc585-29cf-4658-8b97-a23d3e57f349" containerName="registry-server" containerID="cri-o://b2bb98a1b0ef323644a879966667b197d3247567d8f8710e9855a89b5282b901" gracePeriod=2 Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.504215 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btqfx\" (UniqueName: \"kubernetes.io/projected/94dd97ab-9637-4eab-82dc-5d070129abac-kube-api-access-btqfx\") pod \"nmstate-metrics-fdff9cb8d-652tz\" (UID: \"94dd97ab-9637-4eab-82dc-5d070129abac\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-652tz" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.504411 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7thx\" (UniqueName: \"kubernetes.io/projected/b111b867-4d61-489f-aa89-f820b50c9a92-kube-api-access-k7thx\") pod \"nmstate-handler-vknp5\" (UID: \"b111b867-4d61-489f-aa89-f820b50c9a92\") " pod="openshift-nmstate/nmstate-handler-vknp5" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.538779 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-652tz" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.551389 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-fp4rx" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.573898 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a33377f4-f20e-42e4-af02-1f9a7bffc587-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-4xjn2\" (UID: \"a33377f4-f20e-42e4-af02-1f9a7bffc587\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4xjn2" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.573986 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/a33377f4-f20e-42e4-af02-1f9a7bffc587-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-4xjn2\" (UID: \"a33377f4-f20e-42e4-af02-1f9a7bffc587\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4xjn2" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.574008 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2cs2\" (UniqueName: \"kubernetes.io/projected/a33377f4-f20e-42e4-af02-1f9a7bffc587-kube-api-access-k2cs2\") pod \"nmstate-console-plugin-6b874cbd85-4xjn2\" (UID: \"a33377f4-f20e-42e4-af02-1f9a7bffc587\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4xjn2" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.588728 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-vknp5" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.613099 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zlx4b" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.615511 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zlx4b" Oct 03 13:05:31 crc kubenswrapper[4578]: W1003 13:05:31.626349 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb111b867_4d61_489f_aa89_f820b50c9a92.slice/crio-ae2a302e55428ccd1baa687b4ab98e06b22e942f8e29c7b5e37e20133f5d17d9 WatchSource:0}: Error finding container ae2a302e55428ccd1baa687b4ab98e06b22e942f8e29c7b5e37e20133f5d17d9: Status 404 returned error can't find the container with id ae2a302e55428ccd1baa687b4ab98e06b22e942f8e29c7b5e37e20133f5d17d9 Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.675557 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/a33377f4-f20e-42e4-af02-1f9a7bffc587-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-4xjn2\" (UID: \"a33377f4-f20e-42e4-af02-1f9a7bffc587\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4xjn2" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.675604 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2cs2\" (UniqueName: \"kubernetes.io/projected/a33377f4-f20e-42e4-af02-1f9a7bffc587-kube-api-access-k2cs2\") pod \"nmstate-console-plugin-6b874cbd85-4xjn2\" (UID: \"a33377f4-f20e-42e4-af02-1f9a7bffc587\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4xjn2" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.675673 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a33377f4-f20e-42e4-af02-1f9a7bffc587-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-4xjn2\" (UID: \"a33377f4-f20e-42e4-af02-1f9a7bffc587\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4xjn2" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.677279 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/a33377f4-f20e-42e4-af02-1f9a7bffc587-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-4xjn2\" (UID: \"a33377f4-f20e-42e4-af02-1f9a7bffc587\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4xjn2" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.687572 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a33377f4-f20e-42e4-af02-1f9a7bffc587-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-4xjn2\" (UID: \"a33377f4-f20e-42e4-af02-1f9a7bffc587\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4xjn2" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.705381 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7557495f48-jg84f"] Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.706075 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.707441 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2cs2\" (UniqueName: \"kubernetes.io/projected/a33377f4-f20e-42e4-af02-1f9a7bffc587-kube-api-access-k2cs2\") pod \"nmstate-console-plugin-6b874cbd85-4xjn2\" (UID: \"a33377f4-f20e-42e4-af02-1f9a7bffc587\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4xjn2" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.715993 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4xjn2" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.725818 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7557495f48-jg84f"] Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.747174 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zlx4b" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.830012 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-652tz"] Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.877787 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/06817bc1-9882-4a01-bf36-a0ea24509586-console-serving-cert\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.877828 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06817bc1-9882-4a01-bf36-a0ea24509586-trusted-ca-bundle\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.877850 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khgxf\" (UniqueName: \"kubernetes.io/projected/06817bc1-9882-4a01-bf36-a0ea24509586-kube-api-access-khgxf\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.877871 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/06817bc1-9882-4a01-bf36-a0ea24509586-service-ca\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.877893 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/06817bc1-9882-4a01-bf36-a0ea24509586-console-oauth-config\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.877926 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/06817bc1-9882-4a01-bf36-a0ea24509586-console-config\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.877959 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/06817bc1-9882-4a01-bf36-a0ea24509586-oauth-serving-cert\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.978877 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/06817bc1-9882-4a01-bf36-a0ea24509586-oauth-serving-cert\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.978972 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/06817bc1-9882-4a01-bf36-a0ea24509586-console-serving-cert\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.979000 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06817bc1-9882-4a01-bf36-a0ea24509586-trusted-ca-bundle\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.979025 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khgxf\" (UniqueName: \"kubernetes.io/projected/06817bc1-9882-4a01-bf36-a0ea24509586-kube-api-access-khgxf\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.979056 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/06817bc1-9882-4a01-bf36-a0ea24509586-service-ca\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.979083 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/06817bc1-9882-4a01-bf36-a0ea24509586-console-oauth-config\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.979119 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/06817bc1-9882-4a01-bf36-a0ea24509586-console-config\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.980142 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/06817bc1-9882-4a01-bf36-a0ea24509586-console-config\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.980708 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/06817bc1-9882-4a01-bf36-a0ea24509586-oauth-serving-cert\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.981881 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/06817bc1-9882-4a01-bf36-a0ea24509586-service-ca\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.986931 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/06817bc1-9882-4a01-bf36-a0ea24509586-trusted-ca-bundle\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:31 crc kubenswrapper[4578]: I1003 13:05:31.987882 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/06817bc1-9882-4a01-bf36-a0ea24509586-console-serving-cert\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.003187 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/06817bc1-9882-4a01-bf36-a0ea24509586-console-oauth-config\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.006692 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khgxf\" (UniqueName: \"kubernetes.io/projected/06817bc1-9882-4a01-bf36-a0ea24509586-kube-api-access-khgxf\") pod \"console-7557495f48-jg84f\" (UID: \"06817bc1-9882-4a01-bf36-a0ea24509586\") " pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.030847 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-vknp5" event={"ID":"b111b867-4d61-489f-aa89-f820b50c9a92","Type":"ContainerStarted","Data":"ae2a302e55428ccd1baa687b4ab98e06b22e942f8e29c7b5e37e20133f5d17d9"} Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.034354 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-652tz" event={"ID":"94dd97ab-9637-4eab-82dc-5d070129abac","Type":"ContainerStarted","Data":"b995cbc2d550a61c6cb353152755c44f940644cfb453a0753c343505ea136323"} Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.041775 4578 generic.go:334] "Generic (PLEG): container finished" podID="d5afc585-29cf-4658-8b97-a23d3e57f349" containerID="b2bb98a1b0ef323644a879966667b197d3247567d8f8710e9855a89b5282b901" exitCode=0 Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.042534 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2b49" event={"ID":"d5afc585-29cf-4658-8b97-a23d3e57f349","Type":"ContainerDied","Data":"b2bb98a1b0ef323644a879966667b197d3247567d8f8710e9855a89b5282b901"} Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.044786 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x2b49" Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.075939 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.077270 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-4xjn2"] Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.102834 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-fp4rx"] Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.118123 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zlx4b" Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.182769 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62p64\" (UniqueName: \"kubernetes.io/projected/d5afc585-29cf-4658-8b97-a23d3e57f349-kube-api-access-62p64\") pod \"d5afc585-29cf-4658-8b97-a23d3e57f349\" (UID: \"d5afc585-29cf-4658-8b97-a23d3e57f349\") " Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.183118 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5afc585-29cf-4658-8b97-a23d3e57f349-utilities\") pod \"d5afc585-29cf-4658-8b97-a23d3e57f349\" (UID: \"d5afc585-29cf-4658-8b97-a23d3e57f349\") " Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.183180 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5afc585-29cf-4658-8b97-a23d3e57f349-catalog-content\") pod \"d5afc585-29cf-4658-8b97-a23d3e57f349\" (UID: \"d5afc585-29cf-4658-8b97-a23d3e57f349\") " Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.184243 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5afc585-29cf-4658-8b97-a23d3e57f349-utilities" (OuterVolumeSpecName: "utilities") pod "d5afc585-29cf-4658-8b97-a23d3e57f349" (UID: "d5afc585-29cf-4658-8b97-a23d3e57f349"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.185787 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5afc585-29cf-4658-8b97-a23d3e57f349-kube-api-access-62p64" (OuterVolumeSpecName: "kube-api-access-62p64") pod "d5afc585-29cf-4658-8b97-a23d3e57f349" (UID: "d5afc585-29cf-4658-8b97-a23d3e57f349"). InnerVolumeSpecName "kube-api-access-62p64". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.227256 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5afc585-29cf-4658-8b97-a23d3e57f349-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5afc585-29cf-4658-8b97-a23d3e57f349" (UID: "d5afc585-29cf-4658-8b97-a23d3e57f349"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.285334 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5afc585-29cf-4658-8b97-a23d3e57f349-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.285368 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5afc585-29cf-4658-8b97-a23d3e57f349-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.285381 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62p64\" (UniqueName: \"kubernetes.io/projected/d5afc585-29cf-4658-8b97-a23d3e57f349-kube-api-access-62p64\") on node \"crc\" DevicePath \"\"" Oct 03 13:05:32 crc kubenswrapper[4578]: I1003 13:05:32.393404 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7557495f48-jg84f"] Oct 03 13:05:33 crc kubenswrapper[4578]: I1003 13:05:33.055899 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2b49" event={"ID":"d5afc585-29cf-4658-8b97-a23d3e57f349","Type":"ContainerDied","Data":"34e0aff2d27ec0a2d57fe9f346465a58afd08541ca618d46e4316907f78ae660"} Oct 03 13:05:33 crc kubenswrapper[4578]: I1003 13:05:33.056241 4578 scope.go:117] "RemoveContainer" containerID="b2bb98a1b0ef323644a879966667b197d3247567d8f8710e9855a89b5282b901" Oct 03 13:05:33 crc kubenswrapper[4578]: I1003 13:05:33.056133 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x2b49" Oct 03 13:05:33 crc kubenswrapper[4578]: I1003 13:05:33.058864 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4xjn2" event={"ID":"a33377f4-f20e-42e4-af02-1f9a7bffc587","Type":"ContainerStarted","Data":"59f189bb9c70d99f1effd81352e7c359395bf3632c300808646c39864e10bfc7"} Oct 03 13:05:33 crc kubenswrapper[4578]: I1003 13:05:33.068531 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-fp4rx" event={"ID":"bf9f48d4-44a0-4a43-872d-fce41608a68e","Type":"ContainerStarted","Data":"2b9959ba17414c8cbffaeae818cb12e26a81555d9c00cac025432049d6c1bcfc"} Oct 03 13:05:33 crc kubenswrapper[4578]: I1003 13:05:33.074374 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7557495f48-jg84f" event={"ID":"06817bc1-9882-4a01-bf36-a0ea24509586","Type":"ContainerStarted","Data":"d290ae6f1652db4ed2a8e94d8625348c3812324067a867fd5cc1c9adeeda7a6c"} Oct 03 13:05:33 crc kubenswrapper[4578]: I1003 13:05:33.074677 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7557495f48-jg84f" event={"ID":"06817bc1-9882-4a01-bf36-a0ea24509586","Type":"ContainerStarted","Data":"59cb89b112d41becaba1dd7dc68dfbe7a56e66f79b6b5b345846ca7dd0b04ad7"} Oct 03 13:05:33 crc kubenswrapper[4578]: I1003 13:05:33.075102 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x2b49"] Oct 03 13:05:33 crc kubenswrapper[4578]: I1003 13:05:33.083975 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-x2b49"] Oct 03 13:05:33 crc kubenswrapper[4578]: I1003 13:05:33.089146 4578 scope.go:117] "RemoveContainer" containerID="f518b9344b33e05af57e17333d789e978d57f9cc6410d35391a24ec76351ab7f" Oct 03 13:05:33 crc kubenswrapper[4578]: I1003 13:05:33.102174 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7557495f48-jg84f" podStartSLOduration=2.102152069 podStartE2EDuration="2.102152069s" podCreationTimestamp="2025-10-03 13:05:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:05:33.096511599 +0000 UTC m=+868.894983783" watchObservedRunningTime="2025-10-03 13:05:33.102152069 +0000 UTC m=+868.900624253" Oct 03 13:05:33 crc kubenswrapper[4578]: I1003 13:05:33.111365 4578 scope.go:117] "RemoveContainer" containerID="7f9085a5a6652bc39bc85d9e76e78f013cbe4faf491c45f165d5b4ee754787a3" Oct 03 13:05:34 crc kubenswrapper[4578]: I1003 13:05:34.739942 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-b4g4t" Oct 03 13:05:34 crc kubenswrapper[4578]: I1003 13:05:34.805289 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-b4g4t" Oct 03 13:05:34 crc kubenswrapper[4578]: I1003 13:05:34.919858 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5afc585-29cf-4658-8b97-a23d3e57f349" path="/var/lib/kubelet/pods/d5afc585-29cf-4658-8b97-a23d3e57f349/volumes" Oct 03 13:05:35 crc kubenswrapper[4578]: I1003 13:05:35.091838 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:05:35 crc kubenswrapper[4578]: I1003 13:05:35.091883 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:05:35 crc kubenswrapper[4578]: I1003 13:05:35.493085 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zlx4b"] Oct 03 13:05:35 crc kubenswrapper[4578]: I1003 13:05:35.493317 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zlx4b" podUID="15764dfd-829c-45fb-bc6c-7a7e918d6a3c" containerName="registry-server" containerID="cri-o://009ff34b913cdedd4457f95fa360305de69f925ff6d45d342f1c00f3618780eb" gracePeriod=2 Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.045783 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zlx4b" Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.093243 4578 generic.go:334] "Generic (PLEG): container finished" podID="15764dfd-829c-45fb-bc6c-7a7e918d6a3c" containerID="009ff34b913cdedd4457f95fa360305de69f925ff6d45d342f1c00f3618780eb" exitCode=0 Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.093284 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zlx4b" event={"ID":"15764dfd-829c-45fb-bc6c-7a7e918d6a3c","Type":"ContainerDied","Data":"009ff34b913cdedd4457f95fa360305de69f925ff6d45d342f1c00f3618780eb"} Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.093310 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zlx4b" event={"ID":"15764dfd-829c-45fb-bc6c-7a7e918d6a3c","Type":"ContainerDied","Data":"623b28423b49f2b7b9ef5ec2e4f2f96980cbf8ba8fbe453a6ff635c3bb95fed6"} Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.093325 4578 scope.go:117] "RemoveContainer" containerID="009ff34b913cdedd4457f95fa360305de69f925ff6d45d342f1c00f3618780eb" Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.093424 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zlx4b" Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.116288 4578 scope.go:117] "RemoveContainer" containerID="4d9960ead89232c705a3d0e8cb83e4e941fd1e5c72f973c02ffb9350369782b8" Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.134383 4578 scope.go:117] "RemoveContainer" containerID="26787cdd8a79598548808e3c514cdd26cc69104804931dab12c2c02e0c7cddd5" Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.138154 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxn57\" (UniqueName: \"kubernetes.io/projected/15764dfd-829c-45fb-bc6c-7a7e918d6a3c-kube-api-access-kxn57\") pod \"15764dfd-829c-45fb-bc6c-7a7e918d6a3c\" (UID: \"15764dfd-829c-45fb-bc6c-7a7e918d6a3c\") " Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.138369 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15764dfd-829c-45fb-bc6c-7a7e918d6a3c-utilities\") pod \"15764dfd-829c-45fb-bc6c-7a7e918d6a3c\" (UID: \"15764dfd-829c-45fb-bc6c-7a7e918d6a3c\") " Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.138417 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15764dfd-829c-45fb-bc6c-7a7e918d6a3c-catalog-content\") pod \"15764dfd-829c-45fb-bc6c-7a7e918d6a3c\" (UID: \"15764dfd-829c-45fb-bc6c-7a7e918d6a3c\") " Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.139881 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15764dfd-829c-45fb-bc6c-7a7e918d6a3c-utilities" (OuterVolumeSpecName: "utilities") pod "15764dfd-829c-45fb-bc6c-7a7e918d6a3c" (UID: "15764dfd-829c-45fb-bc6c-7a7e918d6a3c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.142242 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15764dfd-829c-45fb-bc6c-7a7e918d6a3c-kube-api-access-kxn57" (OuterVolumeSpecName: "kube-api-access-kxn57") pod "15764dfd-829c-45fb-bc6c-7a7e918d6a3c" (UID: "15764dfd-829c-45fb-bc6c-7a7e918d6a3c"). InnerVolumeSpecName "kube-api-access-kxn57". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.156473 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15764dfd-829c-45fb-bc6c-7a7e918d6a3c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15764dfd-829c-45fb-bc6c-7a7e918d6a3c" (UID: "15764dfd-829c-45fb-bc6c-7a7e918d6a3c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.158014 4578 scope.go:117] "RemoveContainer" containerID="009ff34b913cdedd4457f95fa360305de69f925ff6d45d342f1c00f3618780eb" Oct 03 13:05:36 crc kubenswrapper[4578]: E1003 13:05:36.158488 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"009ff34b913cdedd4457f95fa360305de69f925ff6d45d342f1c00f3618780eb\": container with ID starting with 009ff34b913cdedd4457f95fa360305de69f925ff6d45d342f1c00f3618780eb not found: ID does not exist" containerID="009ff34b913cdedd4457f95fa360305de69f925ff6d45d342f1c00f3618780eb" Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.158540 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"009ff34b913cdedd4457f95fa360305de69f925ff6d45d342f1c00f3618780eb"} err="failed to get container status \"009ff34b913cdedd4457f95fa360305de69f925ff6d45d342f1c00f3618780eb\": rpc error: code = NotFound desc = could not find container \"009ff34b913cdedd4457f95fa360305de69f925ff6d45d342f1c00f3618780eb\": container with ID starting with 009ff34b913cdedd4457f95fa360305de69f925ff6d45d342f1c00f3618780eb not found: ID does not exist" Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.158565 4578 scope.go:117] "RemoveContainer" containerID="4d9960ead89232c705a3d0e8cb83e4e941fd1e5c72f973c02ffb9350369782b8" Oct 03 13:05:36 crc kubenswrapper[4578]: E1003 13:05:36.158837 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d9960ead89232c705a3d0e8cb83e4e941fd1e5c72f973c02ffb9350369782b8\": container with ID starting with 4d9960ead89232c705a3d0e8cb83e4e941fd1e5c72f973c02ffb9350369782b8 not found: ID does not exist" containerID="4d9960ead89232c705a3d0e8cb83e4e941fd1e5c72f973c02ffb9350369782b8" Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.158889 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d9960ead89232c705a3d0e8cb83e4e941fd1e5c72f973c02ffb9350369782b8"} err="failed to get container status \"4d9960ead89232c705a3d0e8cb83e4e941fd1e5c72f973c02ffb9350369782b8\": rpc error: code = NotFound desc = could not find container \"4d9960ead89232c705a3d0e8cb83e4e941fd1e5c72f973c02ffb9350369782b8\": container with ID starting with 4d9960ead89232c705a3d0e8cb83e4e941fd1e5c72f973c02ffb9350369782b8 not found: ID does not exist" Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.158928 4578 scope.go:117] "RemoveContainer" containerID="26787cdd8a79598548808e3c514cdd26cc69104804931dab12c2c02e0c7cddd5" Oct 03 13:05:36 crc kubenswrapper[4578]: E1003 13:05:36.159388 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26787cdd8a79598548808e3c514cdd26cc69104804931dab12c2c02e0c7cddd5\": container with ID starting with 26787cdd8a79598548808e3c514cdd26cc69104804931dab12c2c02e0c7cddd5 not found: ID does not exist" containerID="26787cdd8a79598548808e3c514cdd26cc69104804931dab12c2c02e0c7cddd5" Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.159420 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26787cdd8a79598548808e3c514cdd26cc69104804931dab12c2c02e0c7cddd5"} err="failed to get container status \"26787cdd8a79598548808e3c514cdd26cc69104804931dab12c2c02e0c7cddd5\": rpc error: code = NotFound desc = could not find container \"26787cdd8a79598548808e3c514cdd26cc69104804931dab12c2c02e0c7cddd5\": container with ID starting with 26787cdd8a79598548808e3c514cdd26cc69104804931dab12c2c02e0c7cddd5 not found: ID does not exist" Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.241307 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15764dfd-829c-45fb-bc6c-7a7e918d6a3c-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.241339 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15764dfd-829c-45fb-bc6c-7a7e918d6a3c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.241350 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxn57\" (UniqueName: \"kubernetes.io/projected/15764dfd-829c-45fb-bc6c-7a7e918d6a3c-kube-api-access-kxn57\") on node \"crc\" DevicePath \"\"" Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.419906 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zlx4b"] Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.423922 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zlx4b"] Oct 03 13:05:36 crc kubenswrapper[4578]: I1003 13:05:36.915564 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15764dfd-829c-45fb-bc6c-7a7e918d6a3c" path="/var/lib/kubelet/pods/15764dfd-829c-45fb-bc6c-7a7e918d6a3c/volumes" Oct 03 13:05:37 crc kubenswrapper[4578]: I1003 13:05:37.103038 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-fp4rx" event={"ID":"bf9f48d4-44a0-4a43-872d-fce41608a68e","Type":"ContainerStarted","Data":"98b0dff32f33fc81880bf33f2d1a240442e65e0bf2d988db55129d3890f48d30"} Oct 03 13:05:37 crc kubenswrapper[4578]: I1003 13:05:37.103107 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-fp4rx" Oct 03 13:05:37 crc kubenswrapper[4578]: I1003 13:05:37.112136 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-vknp5" event={"ID":"b111b867-4d61-489f-aa89-f820b50c9a92","Type":"ContainerStarted","Data":"a979aad9900bc495033f15d250dbc223f0f08749dfdda7954ead84cb42090470"} Oct 03 13:05:37 crc kubenswrapper[4578]: I1003 13:05:37.112577 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-vknp5" Oct 03 13:05:37 crc kubenswrapper[4578]: I1003 13:05:37.115007 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-652tz" event={"ID":"94dd97ab-9637-4eab-82dc-5d070129abac","Type":"ContainerStarted","Data":"531c0b8a06f6312edcf4eaacc00d0ced0baaa85c641c98c6ac36b9302c9ff653"} Oct 03 13:05:37 crc kubenswrapper[4578]: I1003 13:05:37.123925 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-fp4rx" podStartSLOduration=2.175445793 podStartE2EDuration="6.123905865s" podCreationTimestamp="2025-10-03 13:05:31 +0000 UTC" firstStartedPulling="2025-10-03 13:05:32.124000635 +0000 UTC m=+867.922472819" lastFinishedPulling="2025-10-03 13:05:36.072460707 +0000 UTC m=+871.870932891" observedRunningTime="2025-10-03 13:05:37.118239964 +0000 UTC m=+872.916712168" watchObservedRunningTime="2025-10-03 13:05:37.123905865 +0000 UTC m=+872.922378049" Oct 03 13:05:37 crc kubenswrapper[4578]: I1003 13:05:37.126670 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4xjn2" event={"ID":"a33377f4-f20e-42e4-af02-1f9a7bffc587","Type":"ContainerStarted","Data":"4a189daee4ebbef6d47ae8c88ae610a6b193bb948f26089f57566999503fc550"} Oct 03 13:05:37 crc kubenswrapper[4578]: I1003 13:05:37.134047 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-vknp5" podStartSLOduration=1.728011345 podStartE2EDuration="6.134028932s" podCreationTimestamp="2025-10-03 13:05:31 +0000 UTC" firstStartedPulling="2025-10-03 13:05:31.641132724 +0000 UTC m=+867.439604908" lastFinishedPulling="2025-10-03 13:05:36.047150311 +0000 UTC m=+871.845622495" observedRunningTime="2025-10-03 13:05:37.133890407 +0000 UTC m=+872.932362591" watchObservedRunningTime="2025-10-03 13:05:37.134028932 +0000 UTC m=+872.932501116" Oct 03 13:05:37 crc kubenswrapper[4578]: I1003 13:05:37.158464 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-4xjn2" podStartSLOduration=2.225588938 podStartE2EDuration="6.158447289s" podCreationTimestamp="2025-10-03 13:05:31 +0000 UTC" firstStartedPulling="2025-10-03 13:05:32.109574678 +0000 UTC m=+867.908046862" lastFinishedPulling="2025-10-03 13:05:36.042433039 +0000 UTC m=+871.840905213" observedRunningTime="2025-10-03 13:05:37.150886551 +0000 UTC m=+872.949358785" watchObservedRunningTime="2025-10-03 13:05:37.158447289 +0000 UTC m=+872.956919473" Oct 03 13:05:38 crc kubenswrapper[4578]: I1003 13:05:38.895381 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b4g4t"] Oct 03 13:05:38 crc kubenswrapper[4578]: I1003 13:05:38.895928 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-b4g4t" podUID="8c92958b-5203-4955-b58a-0869cd4560a5" containerName="registry-server" containerID="cri-o://5d10497f11b7ec48957c604b4a7648ad61c90eac6b368cb8f4a49db6cbd873cc" gracePeriod=2 Oct 03 13:05:39 crc kubenswrapper[4578]: I1003 13:05:39.141075 4578 generic.go:334] "Generic (PLEG): container finished" podID="8c92958b-5203-4955-b58a-0869cd4560a5" containerID="5d10497f11b7ec48957c604b4a7648ad61c90eac6b368cb8f4a49db6cbd873cc" exitCode=0 Oct 03 13:05:39 crc kubenswrapper[4578]: I1003 13:05:39.141141 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b4g4t" event={"ID":"8c92958b-5203-4955-b58a-0869cd4560a5","Type":"ContainerDied","Data":"5d10497f11b7ec48957c604b4a7648ad61c90eac6b368cb8f4a49db6cbd873cc"} Oct 03 13:05:39 crc kubenswrapper[4578]: I1003 13:05:39.142990 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-652tz" event={"ID":"94dd97ab-9637-4eab-82dc-5d070129abac","Type":"ContainerStarted","Data":"1edf8fc49b3414cf5e29a95e2f620e827553815356a1152f578e7037f248ceab"} Oct 03 13:05:39 crc kubenswrapper[4578]: I1003 13:05:39.160539 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-652tz" podStartSLOduration=1.3761259209999999 podStartE2EDuration="8.160521131s" podCreationTimestamp="2025-10-03 13:05:31 +0000 UTC" firstStartedPulling="2025-10-03 13:05:31.841492928 +0000 UTC m=+867.639965112" lastFinishedPulling="2025-10-03 13:05:38.625888138 +0000 UTC m=+874.424360322" observedRunningTime="2025-10-03 13:05:39.158432518 +0000 UTC m=+874.956904712" watchObservedRunningTime="2025-10-03 13:05:39.160521131 +0000 UTC m=+874.958993315" Oct 03 13:05:39 crc kubenswrapper[4578]: I1003 13:05:39.829622 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b4g4t" Oct 03 13:05:39 crc kubenswrapper[4578]: I1003 13:05:39.891165 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c92958b-5203-4955-b58a-0869cd4560a5-catalog-content\") pod \"8c92958b-5203-4955-b58a-0869cd4560a5\" (UID: \"8c92958b-5203-4955-b58a-0869cd4560a5\") " Oct 03 13:05:39 crc kubenswrapper[4578]: I1003 13:05:39.891358 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzm6r\" (UniqueName: \"kubernetes.io/projected/8c92958b-5203-4955-b58a-0869cd4560a5-kube-api-access-pzm6r\") pod \"8c92958b-5203-4955-b58a-0869cd4560a5\" (UID: \"8c92958b-5203-4955-b58a-0869cd4560a5\") " Oct 03 13:05:39 crc kubenswrapper[4578]: I1003 13:05:39.891404 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c92958b-5203-4955-b58a-0869cd4560a5-utilities\") pod \"8c92958b-5203-4955-b58a-0869cd4560a5\" (UID: \"8c92958b-5203-4955-b58a-0869cd4560a5\") " Oct 03 13:05:39 crc kubenswrapper[4578]: I1003 13:05:39.892525 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c92958b-5203-4955-b58a-0869cd4560a5-utilities" (OuterVolumeSpecName: "utilities") pod "8c92958b-5203-4955-b58a-0869cd4560a5" (UID: "8c92958b-5203-4955-b58a-0869cd4560a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:05:39 crc kubenswrapper[4578]: I1003 13:05:39.896169 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c92958b-5203-4955-b58a-0869cd4560a5-kube-api-access-pzm6r" (OuterVolumeSpecName: "kube-api-access-pzm6r") pod "8c92958b-5203-4955-b58a-0869cd4560a5" (UID: "8c92958b-5203-4955-b58a-0869cd4560a5"). InnerVolumeSpecName "kube-api-access-pzm6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:05:39 crc kubenswrapper[4578]: I1003 13:05:39.971645 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c92958b-5203-4955-b58a-0869cd4560a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c92958b-5203-4955-b58a-0869cd4560a5" (UID: "8c92958b-5203-4955-b58a-0869cd4560a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:05:39 crc kubenswrapper[4578]: I1003 13:05:39.993241 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzm6r\" (UniqueName: \"kubernetes.io/projected/8c92958b-5203-4955-b58a-0869cd4560a5-kube-api-access-pzm6r\") on node \"crc\" DevicePath \"\"" Oct 03 13:05:39 crc kubenswrapper[4578]: I1003 13:05:39.993268 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c92958b-5203-4955-b58a-0869cd4560a5-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:05:39 crc kubenswrapper[4578]: I1003 13:05:39.993279 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c92958b-5203-4955-b58a-0869cd4560a5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:05:40 crc kubenswrapper[4578]: I1003 13:05:40.150560 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b4g4t" event={"ID":"8c92958b-5203-4955-b58a-0869cd4560a5","Type":"ContainerDied","Data":"603e768d0128dd051c26180d5adf0cd0f1359210195c871fa19d9b091f5fc702"} Oct 03 13:05:40 crc kubenswrapper[4578]: I1003 13:05:40.150586 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b4g4t" Oct 03 13:05:40 crc kubenswrapper[4578]: I1003 13:05:40.150612 4578 scope.go:117] "RemoveContainer" containerID="5d10497f11b7ec48957c604b4a7648ad61c90eac6b368cb8f4a49db6cbd873cc" Oct 03 13:05:40 crc kubenswrapper[4578]: I1003 13:05:40.167227 4578 scope.go:117] "RemoveContainer" containerID="918a8245b3377a4220b0167bd88f26327bbbe72bddac7fd3d2bba7700bb98ef2" Oct 03 13:05:40 crc kubenswrapper[4578]: I1003 13:05:40.179925 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b4g4t"] Oct 03 13:05:40 crc kubenswrapper[4578]: I1003 13:05:40.182662 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-b4g4t"] Oct 03 13:05:40 crc kubenswrapper[4578]: I1003 13:05:40.211223 4578 scope.go:117] "RemoveContainer" containerID="19608660184ce9b40b6ef8086157aff4e253f41d68ce6a3b8ce52d98d27785f6" Oct 03 13:05:40 crc kubenswrapper[4578]: I1003 13:05:40.918110 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c92958b-5203-4955-b58a-0869cd4560a5" path="/var/lib/kubelet/pods/8c92958b-5203-4955-b58a-0869cd4560a5/volumes" Oct 03 13:05:41 crc kubenswrapper[4578]: I1003 13:05:41.607234 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-vknp5" Oct 03 13:05:42 crc kubenswrapper[4578]: I1003 13:05:42.077214 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:42 crc kubenswrapper[4578]: I1003 13:05:42.077275 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:42 crc kubenswrapper[4578]: I1003 13:05:42.082084 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:42 crc kubenswrapper[4578]: I1003 13:05:42.166911 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7557495f48-jg84f" Oct 03 13:05:42 crc kubenswrapper[4578]: I1003 13:05:42.218171 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-ns26l"] Oct 03 13:05:51 crc kubenswrapper[4578]: I1003 13:05:51.561063 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-fp4rx" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.476870 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj"] Oct 03 13:06:04 crc kubenswrapper[4578]: E1003 13:06:04.477887 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c92958b-5203-4955-b58a-0869cd4560a5" containerName="registry-server" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.477904 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c92958b-5203-4955-b58a-0869cd4560a5" containerName="registry-server" Oct 03 13:06:04 crc kubenswrapper[4578]: E1003 13:06:04.477921 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15764dfd-829c-45fb-bc6c-7a7e918d6a3c" containerName="registry-server" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.477931 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="15764dfd-829c-45fb-bc6c-7a7e918d6a3c" containerName="registry-server" Oct 03 13:06:04 crc kubenswrapper[4578]: E1003 13:06:04.477946 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15764dfd-829c-45fb-bc6c-7a7e918d6a3c" containerName="extract-utilities" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.477955 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="15764dfd-829c-45fb-bc6c-7a7e918d6a3c" containerName="extract-utilities" Oct 03 13:06:04 crc kubenswrapper[4578]: E1003 13:06:04.477969 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c92958b-5203-4955-b58a-0869cd4560a5" containerName="extract-content" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.477977 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c92958b-5203-4955-b58a-0869cd4560a5" containerName="extract-content" Oct 03 13:06:04 crc kubenswrapper[4578]: E1003 13:06:04.477990 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15764dfd-829c-45fb-bc6c-7a7e918d6a3c" containerName="extract-content" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.477999 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="15764dfd-829c-45fb-bc6c-7a7e918d6a3c" containerName="extract-content" Oct 03 13:06:04 crc kubenswrapper[4578]: E1003 13:06:04.478013 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5afc585-29cf-4658-8b97-a23d3e57f349" containerName="extract-utilities" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.478021 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5afc585-29cf-4658-8b97-a23d3e57f349" containerName="extract-utilities" Oct 03 13:06:04 crc kubenswrapper[4578]: E1003 13:06:04.478035 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5afc585-29cf-4658-8b97-a23d3e57f349" containerName="extract-content" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.478044 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5afc585-29cf-4658-8b97-a23d3e57f349" containerName="extract-content" Oct 03 13:06:04 crc kubenswrapper[4578]: E1003 13:06:04.478057 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5afc585-29cf-4658-8b97-a23d3e57f349" containerName="registry-server" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.478066 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5afc585-29cf-4658-8b97-a23d3e57f349" containerName="registry-server" Oct 03 13:06:04 crc kubenswrapper[4578]: E1003 13:06:04.478076 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c92958b-5203-4955-b58a-0869cd4560a5" containerName="extract-utilities" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.478085 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c92958b-5203-4955-b58a-0869cd4560a5" containerName="extract-utilities" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.478224 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c92958b-5203-4955-b58a-0869cd4560a5" containerName="registry-server" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.478245 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="15764dfd-829c-45fb-bc6c-7a7e918d6a3c" containerName="registry-server" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.478262 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5afc585-29cf-4658-8b97-a23d3e57f349" containerName="registry-server" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.479251 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.481643 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.485785 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj"] Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.595913 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzzmr\" (UniqueName: \"kubernetes.io/projected/a16e5036-d08b-4217-b21b-37a28ceb701c-kube-api-access-jzzmr\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj\" (UID: \"a16e5036-d08b-4217-b21b-37a28ceb701c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.595980 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a16e5036-d08b-4217-b21b-37a28ceb701c-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj\" (UID: \"a16e5036-d08b-4217-b21b-37a28ceb701c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.596016 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a16e5036-d08b-4217-b21b-37a28ceb701c-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj\" (UID: \"a16e5036-d08b-4217-b21b-37a28ceb701c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.696903 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzzmr\" (UniqueName: \"kubernetes.io/projected/a16e5036-d08b-4217-b21b-37a28ceb701c-kube-api-access-jzzmr\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj\" (UID: \"a16e5036-d08b-4217-b21b-37a28ceb701c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.696975 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a16e5036-d08b-4217-b21b-37a28ceb701c-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj\" (UID: \"a16e5036-d08b-4217-b21b-37a28ceb701c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.697008 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a16e5036-d08b-4217-b21b-37a28ceb701c-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj\" (UID: \"a16e5036-d08b-4217-b21b-37a28ceb701c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.697436 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a16e5036-d08b-4217-b21b-37a28ceb701c-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj\" (UID: \"a16e5036-d08b-4217-b21b-37a28ceb701c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.697736 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a16e5036-d08b-4217-b21b-37a28ceb701c-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj\" (UID: \"a16e5036-d08b-4217-b21b-37a28ceb701c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.715252 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzzmr\" (UniqueName: \"kubernetes.io/projected/a16e5036-d08b-4217-b21b-37a28ceb701c-kube-api-access-jzzmr\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj\" (UID: \"a16e5036-d08b-4217-b21b-37a28ceb701c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj" Oct 03 13:06:04 crc kubenswrapper[4578]: I1003 13:06:04.795808 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj" Oct 03 13:06:05 crc kubenswrapper[4578]: I1003 13:06:05.092118 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:06:05 crc kubenswrapper[4578]: I1003 13:06:05.092456 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:06:05 crc kubenswrapper[4578]: I1003 13:06:05.183855 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj"] Oct 03 13:06:05 crc kubenswrapper[4578]: I1003 13:06:05.291178 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj" event={"ID":"a16e5036-d08b-4217-b21b-37a28ceb701c","Type":"ContainerStarted","Data":"12ca6a6368b6d017303bc35b6639933e8aa867ab344da0aaaa9b481312974db3"} Oct 03 13:06:06 crc kubenswrapper[4578]: I1003 13:06:06.297196 4578 generic.go:334] "Generic (PLEG): container finished" podID="a16e5036-d08b-4217-b21b-37a28ceb701c" containerID="e9b287098833119a0282a8087ef2a3cd107e0afcdf72cbc0b2606ee1b058a1be" exitCode=0 Oct 03 13:06:06 crc kubenswrapper[4578]: I1003 13:06:06.297391 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj" event={"ID":"a16e5036-d08b-4217-b21b-37a28ceb701c","Type":"ContainerDied","Data":"e9b287098833119a0282a8087ef2a3cd107e0afcdf72cbc0b2606ee1b058a1be"} Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.290104 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-ns26l" podUID="7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6" containerName="console" containerID="cri-o://841a71102e1dc33e360fa2d6d0e2bdb8b21bed4852a9dbaa4e6c2519e39b7a0f" gracePeriod=15 Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.612125 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-ns26l_7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6/console/0.log" Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.612443 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ns26l" Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.740572 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-console-config\") pod \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.740687 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjfxr\" (UniqueName: \"kubernetes.io/projected/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-kube-api-access-bjfxr\") pod \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.740841 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-trusted-ca-bundle\") pod \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.740932 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-console-oauth-config\") pod \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.740991 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-console-serving-cert\") pod \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.741012 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-oauth-serving-cert\") pod \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.741029 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-service-ca\") pod \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\" (UID: \"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6\") " Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.741456 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6" (UID: "7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.741614 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-console-config" (OuterVolumeSpecName: "console-config") pod "7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6" (UID: "7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.741652 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6" (UID: "7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.741693 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-service-ca" (OuterVolumeSpecName: "service-ca") pod "7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6" (UID: "7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.746642 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-kube-api-access-bjfxr" (OuterVolumeSpecName: "kube-api-access-bjfxr") pod "7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6" (UID: "7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6"). InnerVolumeSpecName "kube-api-access-bjfxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.746650 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6" (UID: "7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.746958 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6" (UID: "7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.843011 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjfxr\" (UniqueName: \"kubernetes.io/projected/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-kube-api-access-bjfxr\") on node \"crc\" DevicePath \"\"" Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.843052 4578 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.843068 4578 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.843079 4578 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.843093 4578 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.843104 4578 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-service-ca\") on node \"crc\" DevicePath \"\"" Oct 03 13:06:07 crc kubenswrapper[4578]: I1003 13:06:07.843116 4578 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6-console-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:06:08 crc kubenswrapper[4578]: I1003 13:06:08.310361 4578 generic.go:334] "Generic (PLEG): container finished" podID="a16e5036-d08b-4217-b21b-37a28ceb701c" containerID="24238233b55952554e29a561834c2ba5b0e12919af5f7ee1dfc37fdcbf93e6bd" exitCode=0 Oct 03 13:06:08 crc kubenswrapper[4578]: I1003 13:06:08.310740 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj" event={"ID":"a16e5036-d08b-4217-b21b-37a28ceb701c","Type":"ContainerDied","Data":"24238233b55952554e29a561834c2ba5b0e12919af5f7ee1dfc37fdcbf93e6bd"} Oct 03 13:06:08 crc kubenswrapper[4578]: I1003 13:06:08.312467 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-ns26l_7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6/console/0.log" Oct 03 13:06:08 crc kubenswrapper[4578]: I1003 13:06:08.312521 4578 generic.go:334] "Generic (PLEG): container finished" podID="7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6" containerID="841a71102e1dc33e360fa2d6d0e2bdb8b21bed4852a9dbaa4e6c2519e39b7a0f" exitCode=2 Oct 03 13:06:08 crc kubenswrapper[4578]: I1003 13:06:08.312553 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ns26l" event={"ID":"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6","Type":"ContainerDied","Data":"841a71102e1dc33e360fa2d6d0e2bdb8b21bed4852a9dbaa4e6c2519e39b7a0f"} Oct 03 13:06:08 crc kubenswrapper[4578]: I1003 13:06:08.312563 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ns26l" Oct 03 13:06:08 crc kubenswrapper[4578]: I1003 13:06:08.312582 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ns26l" event={"ID":"7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6","Type":"ContainerDied","Data":"abb6d94d578781509f9906a23838695125978d11ac4cb25e065f03fe458ae071"} Oct 03 13:06:08 crc kubenswrapper[4578]: I1003 13:06:08.312601 4578 scope.go:117] "RemoveContainer" containerID="841a71102e1dc33e360fa2d6d0e2bdb8b21bed4852a9dbaa4e6c2519e39b7a0f" Oct 03 13:06:08 crc kubenswrapper[4578]: I1003 13:06:08.348169 4578 scope.go:117] "RemoveContainer" containerID="841a71102e1dc33e360fa2d6d0e2bdb8b21bed4852a9dbaa4e6c2519e39b7a0f" Oct 03 13:06:08 crc kubenswrapper[4578]: E1003 13:06:08.348574 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"841a71102e1dc33e360fa2d6d0e2bdb8b21bed4852a9dbaa4e6c2519e39b7a0f\": container with ID starting with 841a71102e1dc33e360fa2d6d0e2bdb8b21bed4852a9dbaa4e6c2519e39b7a0f not found: ID does not exist" containerID="841a71102e1dc33e360fa2d6d0e2bdb8b21bed4852a9dbaa4e6c2519e39b7a0f" Oct 03 13:06:08 crc kubenswrapper[4578]: I1003 13:06:08.348602 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"841a71102e1dc33e360fa2d6d0e2bdb8b21bed4852a9dbaa4e6c2519e39b7a0f"} err="failed to get container status \"841a71102e1dc33e360fa2d6d0e2bdb8b21bed4852a9dbaa4e6c2519e39b7a0f\": rpc error: code = NotFound desc = could not find container \"841a71102e1dc33e360fa2d6d0e2bdb8b21bed4852a9dbaa4e6c2519e39b7a0f\": container with ID starting with 841a71102e1dc33e360fa2d6d0e2bdb8b21bed4852a9dbaa4e6c2519e39b7a0f not found: ID does not exist" Oct 03 13:06:08 crc kubenswrapper[4578]: I1003 13:06:08.348769 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-ns26l"] Oct 03 13:06:08 crc kubenswrapper[4578]: I1003 13:06:08.351982 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-ns26l"] Oct 03 13:06:08 crc kubenswrapper[4578]: I1003 13:06:08.914942 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6" path="/var/lib/kubelet/pods/7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6/volumes" Oct 03 13:06:09 crc kubenswrapper[4578]: I1003 13:06:09.322159 4578 generic.go:334] "Generic (PLEG): container finished" podID="a16e5036-d08b-4217-b21b-37a28ceb701c" containerID="08a183a43cc3cad13f70db998442798b22e982d8f6e6f9af60a01055f865b167" exitCode=0 Oct 03 13:06:09 crc kubenswrapper[4578]: I1003 13:06:09.322220 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj" event={"ID":"a16e5036-d08b-4217-b21b-37a28ceb701c","Type":"ContainerDied","Data":"08a183a43cc3cad13f70db998442798b22e982d8f6e6f9af60a01055f865b167"} Oct 03 13:06:10 crc kubenswrapper[4578]: I1003 13:06:10.577369 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj" Oct 03 13:06:10 crc kubenswrapper[4578]: I1003 13:06:10.685994 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzzmr\" (UniqueName: \"kubernetes.io/projected/a16e5036-d08b-4217-b21b-37a28ceb701c-kube-api-access-jzzmr\") pod \"a16e5036-d08b-4217-b21b-37a28ceb701c\" (UID: \"a16e5036-d08b-4217-b21b-37a28ceb701c\") " Oct 03 13:06:10 crc kubenswrapper[4578]: I1003 13:06:10.686085 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a16e5036-d08b-4217-b21b-37a28ceb701c-util\") pod \"a16e5036-d08b-4217-b21b-37a28ceb701c\" (UID: \"a16e5036-d08b-4217-b21b-37a28ceb701c\") " Oct 03 13:06:10 crc kubenswrapper[4578]: I1003 13:06:10.686115 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a16e5036-d08b-4217-b21b-37a28ceb701c-bundle\") pod \"a16e5036-d08b-4217-b21b-37a28ceb701c\" (UID: \"a16e5036-d08b-4217-b21b-37a28ceb701c\") " Oct 03 13:06:10 crc kubenswrapper[4578]: I1003 13:06:10.687068 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a16e5036-d08b-4217-b21b-37a28ceb701c-bundle" (OuterVolumeSpecName: "bundle") pod "a16e5036-d08b-4217-b21b-37a28ceb701c" (UID: "a16e5036-d08b-4217-b21b-37a28ceb701c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:06:10 crc kubenswrapper[4578]: I1003 13:06:10.690805 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a16e5036-d08b-4217-b21b-37a28ceb701c-kube-api-access-jzzmr" (OuterVolumeSpecName: "kube-api-access-jzzmr") pod "a16e5036-d08b-4217-b21b-37a28ceb701c" (UID: "a16e5036-d08b-4217-b21b-37a28ceb701c"). InnerVolumeSpecName "kube-api-access-jzzmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:06:10 crc kubenswrapper[4578]: I1003 13:06:10.788268 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzzmr\" (UniqueName: \"kubernetes.io/projected/a16e5036-d08b-4217-b21b-37a28ceb701c-kube-api-access-jzzmr\") on node \"crc\" DevicePath \"\"" Oct 03 13:06:10 crc kubenswrapper[4578]: I1003 13:06:10.788303 4578 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a16e5036-d08b-4217-b21b-37a28ceb701c-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:06:10 crc kubenswrapper[4578]: I1003 13:06:10.900789 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a16e5036-d08b-4217-b21b-37a28ceb701c-util" (OuterVolumeSpecName: "util") pod "a16e5036-d08b-4217-b21b-37a28ceb701c" (UID: "a16e5036-d08b-4217-b21b-37a28ceb701c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:06:10 crc kubenswrapper[4578]: I1003 13:06:10.991055 4578 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a16e5036-d08b-4217-b21b-37a28ceb701c-util\") on node \"crc\" DevicePath \"\"" Oct 03 13:06:11 crc kubenswrapper[4578]: I1003 13:06:11.335000 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj" event={"ID":"a16e5036-d08b-4217-b21b-37a28ceb701c","Type":"ContainerDied","Data":"12ca6a6368b6d017303bc35b6639933e8aa867ab344da0aaaa9b481312974db3"} Oct 03 13:06:11 crc kubenswrapper[4578]: I1003 13:06:11.335286 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12ca6a6368b6d017303bc35b6639933e8aa867ab344da0aaaa9b481312974db3" Oct 03 13:06:11 crc kubenswrapper[4578]: I1003 13:06:11.335219 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj" Oct 03 13:06:19 crc kubenswrapper[4578]: I1003 13:06:19.774212 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7969464d44-s7mk5"] Oct 03 13:06:19 crc kubenswrapper[4578]: E1003 13:06:19.775137 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6" containerName="console" Oct 03 13:06:19 crc kubenswrapper[4578]: I1003 13:06:19.775152 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6" containerName="console" Oct 03 13:06:19 crc kubenswrapper[4578]: E1003 13:06:19.775170 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a16e5036-d08b-4217-b21b-37a28ceb701c" containerName="util" Oct 03 13:06:19 crc kubenswrapper[4578]: I1003 13:06:19.775176 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a16e5036-d08b-4217-b21b-37a28ceb701c" containerName="util" Oct 03 13:06:19 crc kubenswrapper[4578]: E1003 13:06:19.775189 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a16e5036-d08b-4217-b21b-37a28ceb701c" containerName="extract" Oct 03 13:06:19 crc kubenswrapper[4578]: I1003 13:06:19.775194 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a16e5036-d08b-4217-b21b-37a28ceb701c" containerName="extract" Oct 03 13:06:19 crc kubenswrapper[4578]: E1003 13:06:19.775205 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a16e5036-d08b-4217-b21b-37a28ceb701c" containerName="pull" Oct 03 13:06:19 crc kubenswrapper[4578]: I1003 13:06:19.775211 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a16e5036-d08b-4217-b21b-37a28ceb701c" containerName="pull" Oct 03 13:06:19 crc kubenswrapper[4578]: I1003 13:06:19.775306 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c25fbcb-60e6-4b80-9b5b-6ce8af2d67a6" containerName="console" Oct 03 13:06:19 crc kubenswrapper[4578]: I1003 13:06:19.775323 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="a16e5036-d08b-4217-b21b-37a28ceb701c" containerName="extract" Oct 03 13:06:19 crc kubenswrapper[4578]: I1003 13:06:19.775702 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7969464d44-s7mk5" Oct 03 13:06:19 crc kubenswrapper[4578]: I1003 13:06:19.780861 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 03 13:06:19 crc kubenswrapper[4578]: I1003 13:06:19.781016 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 03 13:06:19 crc kubenswrapper[4578]: I1003 13:06:19.781126 4578 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 03 13:06:19 crc kubenswrapper[4578]: I1003 13:06:19.781360 4578 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-9d62n" Oct 03 13:06:19 crc kubenswrapper[4578]: I1003 13:06:19.781073 4578 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 03 13:06:19 crc kubenswrapper[4578]: I1003 13:06:19.797669 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7969464d44-s7mk5"] Oct 03 13:06:19 crc kubenswrapper[4578]: I1003 13:06:19.895946 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ce9289ad-1745-49b0-af1b-e37b45c999fa-apiservice-cert\") pod \"metallb-operator-controller-manager-7969464d44-s7mk5\" (UID: \"ce9289ad-1745-49b0-af1b-e37b45c999fa\") " pod="metallb-system/metallb-operator-controller-manager-7969464d44-s7mk5" Oct 03 13:06:19 crc kubenswrapper[4578]: I1003 13:06:19.896033 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k89dz\" (UniqueName: \"kubernetes.io/projected/ce9289ad-1745-49b0-af1b-e37b45c999fa-kube-api-access-k89dz\") pod \"metallb-operator-controller-manager-7969464d44-s7mk5\" (UID: \"ce9289ad-1745-49b0-af1b-e37b45c999fa\") " pod="metallb-system/metallb-operator-controller-manager-7969464d44-s7mk5" Oct 03 13:06:19 crc kubenswrapper[4578]: I1003 13:06:19.896072 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ce9289ad-1745-49b0-af1b-e37b45c999fa-webhook-cert\") pod \"metallb-operator-controller-manager-7969464d44-s7mk5\" (UID: \"ce9289ad-1745-49b0-af1b-e37b45c999fa\") " pod="metallb-system/metallb-operator-controller-manager-7969464d44-s7mk5" Oct 03 13:06:19 crc kubenswrapper[4578]: I1003 13:06:19.997523 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ce9289ad-1745-49b0-af1b-e37b45c999fa-apiservice-cert\") pod \"metallb-operator-controller-manager-7969464d44-s7mk5\" (UID: \"ce9289ad-1745-49b0-af1b-e37b45c999fa\") " pod="metallb-system/metallb-operator-controller-manager-7969464d44-s7mk5" Oct 03 13:06:19 crc kubenswrapper[4578]: I1003 13:06:19.997901 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k89dz\" (UniqueName: \"kubernetes.io/projected/ce9289ad-1745-49b0-af1b-e37b45c999fa-kube-api-access-k89dz\") pod \"metallb-operator-controller-manager-7969464d44-s7mk5\" (UID: \"ce9289ad-1745-49b0-af1b-e37b45c999fa\") " pod="metallb-system/metallb-operator-controller-manager-7969464d44-s7mk5" Oct 03 13:06:19 crc kubenswrapper[4578]: I1003 13:06:19.998341 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ce9289ad-1745-49b0-af1b-e37b45c999fa-webhook-cert\") pod \"metallb-operator-controller-manager-7969464d44-s7mk5\" (UID: \"ce9289ad-1745-49b0-af1b-e37b45c999fa\") " pod="metallb-system/metallb-operator-controller-manager-7969464d44-s7mk5" Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.004490 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ce9289ad-1745-49b0-af1b-e37b45c999fa-apiservice-cert\") pod \"metallb-operator-controller-manager-7969464d44-s7mk5\" (UID: \"ce9289ad-1745-49b0-af1b-e37b45c999fa\") " pod="metallb-system/metallb-operator-controller-manager-7969464d44-s7mk5" Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.010282 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ce9289ad-1745-49b0-af1b-e37b45c999fa-webhook-cert\") pod \"metallb-operator-controller-manager-7969464d44-s7mk5\" (UID: \"ce9289ad-1745-49b0-af1b-e37b45c999fa\") " pod="metallb-system/metallb-operator-controller-manager-7969464d44-s7mk5" Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.017127 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k89dz\" (UniqueName: \"kubernetes.io/projected/ce9289ad-1745-49b0-af1b-e37b45c999fa-kube-api-access-k89dz\") pod \"metallb-operator-controller-manager-7969464d44-s7mk5\" (UID: \"ce9289ad-1745-49b0-af1b-e37b45c999fa\") " pod="metallb-system/metallb-operator-controller-manager-7969464d44-s7mk5" Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.090418 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7969464d44-s7mk5" Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.118753 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7b7dc59bdd-xqgvb"] Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.119503 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7b7dc59bdd-xqgvb" Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.123321 4578 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-2mbff" Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.123473 4578 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.123567 4578 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.130574 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7b7dc59bdd-xqgvb"] Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.215567 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/df783eb1-661a-455b-94b2-4045de8dadc0-apiservice-cert\") pod \"metallb-operator-webhook-server-7b7dc59bdd-xqgvb\" (UID: \"df783eb1-661a-455b-94b2-4045de8dadc0\") " pod="metallb-system/metallb-operator-webhook-server-7b7dc59bdd-xqgvb" Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.215610 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/df783eb1-661a-455b-94b2-4045de8dadc0-webhook-cert\") pod \"metallb-operator-webhook-server-7b7dc59bdd-xqgvb\" (UID: \"df783eb1-661a-455b-94b2-4045de8dadc0\") " pod="metallb-system/metallb-operator-webhook-server-7b7dc59bdd-xqgvb" Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.215666 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7bz7\" (UniqueName: \"kubernetes.io/projected/df783eb1-661a-455b-94b2-4045de8dadc0-kube-api-access-j7bz7\") pod \"metallb-operator-webhook-server-7b7dc59bdd-xqgvb\" (UID: \"df783eb1-661a-455b-94b2-4045de8dadc0\") " pod="metallb-system/metallb-operator-webhook-server-7b7dc59bdd-xqgvb" Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.316349 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/df783eb1-661a-455b-94b2-4045de8dadc0-apiservice-cert\") pod \"metallb-operator-webhook-server-7b7dc59bdd-xqgvb\" (UID: \"df783eb1-661a-455b-94b2-4045de8dadc0\") " pod="metallb-system/metallb-operator-webhook-server-7b7dc59bdd-xqgvb" Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.316699 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/df783eb1-661a-455b-94b2-4045de8dadc0-webhook-cert\") pod \"metallb-operator-webhook-server-7b7dc59bdd-xqgvb\" (UID: \"df783eb1-661a-455b-94b2-4045de8dadc0\") " pod="metallb-system/metallb-operator-webhook-server-7b7dc59bdd-xqgvb" Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.316744 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7bz7\" (UniqueName: \"kubernetes.io/projected/df783eb1-661a-455b-94b2-4045de8dadc0-kube-api-access-j7bz7\") pod \"metallb-operator-webhook-server-7b7dc59bdd-xqgvb\" (UID: \"df783eb1-661a-455b-94b2-4045de8dadc0\") " pod="metallb-system/metallb-operator-webhook-server-7b7dc59bdd-xqgvb" Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.325961 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/df783eb1-661a-455b-94b2-4045de8dadc0-apiservice-cert\") pod \"metallb-operator-webhook-server-7b7dc59bdd-xqgvb\" (UID: \"df783eb1-661a-455b-94b2-4045de8dadc0\") " pod="metallb-system/metallb-operator-webhook-server-7b7dc59bdd-xqgvb" Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.328687 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/df783eb1-661a-455b-94b2-4045de8dadc0-webhook-cert\") pod \"metallb-operator-webhook-server-7b7dc59bdd-xqgvb\" (UID: \"df783eb1-661a-455b-94b2-4045de8dadc0\") " pod="metallb-system/metallb-operator-webhook-server-7b7dc59bdd-xqgvb" Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.355048 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7bz7\" (UniqueName: \"kubernetes.io/projected/df783eb1-661a-455b-94b2-4045de8dadc0-kube-api-access-j7bz7\") pod \"metallb-operator-webhook-server-7b7dc59bdd-xqgvb\" (UID: \"df783eb1-661a-455b-94b2-4045de8dadc0\") " pod="metallb-system/metallb-operator-webhook-server-7b7dc59bdd-xqgvb" Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.454793 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7969464d44-s7mk5"] Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.461301 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7b7dc59bdd-xqgvb" Oct 03 13:06:20 crc kubenswrapper[4578]: I1003 13:06:20.820951 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7b7dc59bdd-xqgvb"] Oct 03 13:06:21 crc kubenswrapper[4578]: I1003 13:06:21.385430 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7b7dc59bdd-xqgvb" event={"ID":"df783eb1-661a-455b-94b2-4045de8dadc0","Type":"ContainerStarted","Data":"c36607e34d7adb94ebd021f0ab1982d129fa7a177decc625a5c4e060ea6d8b7d"} Oct 03 13:06:21 crc kubenswrapper[4578]: I1003 13:06:21.386469 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7969464d44-s7mk5" event={"ID":"ce9289ad-1745-49b0-af1b-e37b45c999fa","Type":"ContainerStarted","Data":"f641244d5a6361efdba7e4549fbd1bf9ae322bff712a80d19fbfef0cf2b9d540"} Oct 03 13:06:28 crc kubenswrapper[4578]: I1003 13:06:28.458074 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7969464d44-s7mk5" event={"ID":"ce9289ad-1745-49b0-af1b-e37b45c999fa","Type":"ContainerStarted","Data":"0f7eec17ebe15e39e29203429628b52e261c2e0cfbd0d82fda297b6247bf24ec"} Oct 03 13:06:28 crc kubenswrapper[4578]: I1003 13:06:28.458673 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7969464d44-s7mk5" Oct 03 13:06:28 crc kubenswrapper[4578]: I1003 13:06:28.461129 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7b7dc59bdd-xqgvb" event={"ID":"df783eb1-661a-455b-94b2-4045de8dadc0","Type":"ContainerStarted","Data":"458fa1ab4521cf0523715667ffa5d862af7aeacf021229c2c5389c61c60d3409"} Oct 03 13:06:28 crc kubenswrapper[4578]: I1003 13:06:28.461496 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7b7dc59bdd-xqgvb" Oct 03 13:06:28 crc kubenswrapper[4578]: I1003 13:06:28.486940 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7969464d44-s7mk5" podStartSLOduration=2.629899511 podStartE2EDuration="9.486912539s" podCreationTimestamp="2025-10-03 13:06:19 +0000 UTC" firstStartedPulling="2025-10-03 13:06:20.470895913 +0000 UTC m=+916.269368097" lastFinishedPulling="2025-10-03 13:06:27.327908931 +0000 UTC m=+923.126381125" observedRunningTime="2025-10-03 13:06:28.484052459 +0000 UTC m=+924.282524663" watchObservedRunningTime="2025-10-03 13:06:28.486912539 +0000 UTC m=+924.285384723" Oct 03 13:06:28 crc kubenswrapper[4578]: I1003 13:06:28.517919 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7b7dc59bdd-xqgvb" podStartSLOduration=2.013324294 podStartE2EDuration="8.517898026s" podCreationTimestamp="2025-10-03 13:06:20 +0000 UTC" firstStartedPulling="2025-10-03 13:06:20.831481057 +0000 UTC m=+916.629953231" lastFinishedPulling="2025-10-03 13:06:27.336054779 +0000 UTC m=+923.134526963" observedRunningTime="2025-10-03 13:06:28.51293337 +0000 UTC m=+924.311405554" watchObservedRunningTime="2025-10-03 13:06:28.517898026 +0000 UTC m=+924.316370210" Oct 03 13:06:35 crc kubenswrapper[4578]: I1003 13:06:35.091787 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:06:35 crc kubenswrapper[4578]: I1003 13:06:35.092393 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:06:35 crc kubenswrapper[4578]: I1003 13:06:35.092447 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 13:06:35 crc kubenswrapper[4578]: I1003 13:06:35.093124 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"942bca54ab772aabe22d68921c64f6de2fc3db7407fb9842d79fce2fa9613a96"} pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:06:35 crc kubenswrapper[4578]: I1003 13:06:35.093191 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" containerID="cri-o://942bca54ab772aabe22d68921c64f6de2fc3db7407fb9842d79fce2fa9613a96" gracePeriod=600 Oct 03 13:06:35 crc kubenswrapper[4578]: I1003 13:06:35.498972 4578 generic.go:334] "Generic (PLEG): container finished" podID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerID="942bca54ab772aabe22d68921c64f6de2fc3db7407fb9842d79fce2fa9613a96" exitCode=0 Oct 03 13:06:35 crc kubenswrapper[4578]: I1003 13:06:35.499084 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerDied","Data":"942bca54ab772aabe22d68921c64f6de2fc3db7407fb9842d79fce2fa9613a96"} Oct 03 13:06:35 crc kubenswrapper[4578]: I1003 13:06:35.499382 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"415155e4360685a3545a1494ed131d83906f75047c2e61272798309c3d089d8b"} Oct 03 13:06:35 crc kubenswrapper[4578]: I1003 13:06:35.499408 4578 scope.go:117] "RemoveContainer" containerID="9fca9ac685f413e21ab5da95ad3fbaecf2db649af27947616da386bcc758120f" Oct 03 13:06:40 crc kubenswrapper[4578]: I1003 13:06:40.467267 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7b7dc59bdd-xqgvb" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.092561 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7969464d44-s7mk5" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.852130 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-vcftq"] Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.854716 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.860610 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.864739 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-t7dz4"] Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.865795 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t7dz4" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.866562 4578 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-jnjc7" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.869200 4578 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.876734 4578 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.900604 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-t7dz4"] Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.937380 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f87kv\" (UniqueName: \"kubernetes.io/projected/70df60c7-212d-4834-b237-527ca5710600-kube-api-access-f87kv\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.937427 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/70df60c7-212d-4834-b237-527ca5710600-frr-startup\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.937449 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/70df60c7-212d-4834-b237-527ca5710600-reloader\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.937465 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/70df60c7-212d-4834-b237-527ca5710600-frr-conf\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.937486 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/70df60c7-212d-4834-b237-527ca5710600-metrics\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.937586 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/70df60c7-212d-4834-b237-527ca5710600-metrics-certs\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.937640 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxtxp\" (UniqueName: \"kubernetes.io/projected/3d7fbaf2-ba49-481f-a56c-df68ef840423-kube-api-access-dxtxp\") pod \"frr-k8s-webhook-server-64bf5d555-t7dz4\" (UID: \"3d7fbaf2-ba49-481f-a56c-df68ef840423\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t7dz4" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.937681 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3d7fbaf2-ba49-481f-a56c-df68ef840423-cert\") pod \"frr-k8s-webhook-server-64bf5d555-t7dz4\" (UID: \"3d7fbaf2-ba49-481f-a56c-df68ef840423\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t7dz4" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.937698 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/70df60c7-212d-4834-b237-527ca5710600-frr-sockets\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.969428 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-9sc2x"] Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.970617 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-9sc2x" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.973222 4578 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.973646 4578 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-n92tk" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.973793 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.973877 4578 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.983430 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-jhpgk"] Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.984229 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-jhpgk" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.987038 4578 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 03 13:07:00 crc kubenswrapper[4578]: I1003 13:07:00.999364 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-jhpgk"] Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.039893 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c036c09c-bcdf-4fcd-908a-4f59ff7de9d9-metallb-excludel2\") pod \"speaker-9sc2x\" (UID: \"c036c09c-bcdf-4fcd-908a-4f59ff7de9d9\") " pod="metallb-system/speaker-9sc2x" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.039985 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3d7fbaf2-ba49-481f-a56c-df68ef840423-cert\") pod \"frr-k8s-webhook-server-64bf5d555-t7dz4\" (UID: \"3d7fbaf2-ba49-481f-a56c-df68ef840423\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t7dz4" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.040022 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/70df60c7-212d-4834-b237-527ca5710600-frr-sockets\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.040102 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f87kv\" (UniqueName: \"kubernetes.io/projected/70df60c7-212d-4834-b237-527ca5710600-kube-api-access-f87kv\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.040172 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/70df60c7-212d-4834-b237-527ca5710600-frr-startup\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.040210 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/70df60c7-212d-4834-b237-527ca5710600-reloader\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.040233 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/70df60c7-212d-4834-b237-527ca5710600-frr-conf\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.040296 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/70df60c7-212d-4834-b237-527ca5710600-metrics\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.040346 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/70df60c7-212d-4834-b237-527ca5710600-metrics-certs\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.040371 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c036c09c-bcdf-4fcd-908a-4f59ff7de9d9-memberlist\") pod \"speaker-9sc2x\" (UID: \"c036c09c-bcdf-4fcd-908a-4f59ff7de9d9\") " pod="metallb-system/speaker-9sc2x" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.040407 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxtxp\" (UniqueName: \"kubernetes.io/projected/3d7fbaf2-ba49-481f-a56c-df68ef840423-kube-api-access-dxtxp\") pod \"frr-k8s-webhook-server-64bf5d555-t7dz4\" (UID: \"3d7fbaf2-ba49-481f-a56c-df68ef840423\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t7dz4" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.040434 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jpxf\" (UniqueName: \"kubernetes.io/projected/c036c09c-bcdf-4fcd-908a-4f59ff7de9d9-kube-api-access-4jpxf\") pod \"speaker-9sc2x\" (UID: \"c036c09c-bcdf-4fcd-908a-4f59ff7de9d9\") " pod="metallb-system/speaker-9sc2x" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.040488 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c036c09c-bcdf-4fcd-908a-4f59ff7de9d9-metrics-certs\") pod \"speaker-9sc2x\" (UID: \"c036c09c-bcdf-4fcd-908a-4f59ff7de9d9\") " pod="metallb-system/speaker-9sc2x" Oct 03 13:07:01 crc kubenswrapper[4578]: E1003 13:07:01.040864 4578 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 03 13:07:01 crc kubenswrapper[4578]: E1003 13:07:01.040926 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d7fbaf2-ba49-481f-a56c-df68ef840423-cert podName:3d7fbaf2-ba49-481f-a56c-df68ef840423 nodeName:}" failed. No retries permitted until 2025-10-03 13:07:01.540905587 +0000 UTC m=+957.339377771 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3d7fbaf2-ba49-481f-a56c-df68ef840423-cert") pod "frr-k8s-webhook-server-64bf5d555-t7dz4" (UID: "3d7fbaf2-ba49-481f-a56c-df68ef840423") : secret "frr-k8s-webhook-server-cert" not found Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.041574 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/70df60c7-212d-4834-b237-527ca5710600-frr-sockets\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.043981 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/70df60c7-212d-4834-b237-527ca5710600-frr-startup\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.044229 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/70df60c7-212d-4834-b237-527ca5710600-reloader\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.044463 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/70df60c7-212d-4834-b237-527ca5710600-frr-conf\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.045662 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/70df60c7-212d-4834-b237-527ca5710600-metrics\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.052095 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/70df60c7-212d-4834-b237-527ca5710600-metrics-certs\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.063768 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f87kv\" (UniqueName: \"kubernetes.io/projected/70df60c7-212d-4834-b237-527ca5710600-kube-api-access-f87kv\") pod \"frr-k8s-vcftq\" (UID: \"70df60c7-212d-4834-b237-527ca5710600\") " pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.077319 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxtxp\" (UniqueName: \"kubernetes.io/projected/3d7fbaf2-ba49-481f-a56c-df68ef840423-kube-api-access-dxtxp\") pod \"frr-k8s-webhook-server-64bf5d555-t7dz4\" (UID: \"3d7fbaf2-ba49-481f-a56c-df68ef840423\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t7dz4" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.142022 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c036c09c-bcdf-4fcd-908a-4f59ff7de9d9-memberlist\") pod \"speaker-9sc2x\" (UID: \"c036c09c-bcdf-4fcd-908a-4f59ff7de9d9\") " pod="metallb-system/speaker-9sc2x" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.142073 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jpxf\" (UniqueName: \"kubernetes.io/projected/c036c09c-bcdf-4fcd-908a-4f59ff7de9d9-kube-api-access-4jpxf\") pod \"speaker-9sc2x\" (UID: \"c036c09c-bcdf-4fcd-908a-4f59ff7de9d9\") " pod="metallb-system/speaker-9sc2x" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.142104 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c036c09c-bcdf-4fcd-908a-4f59ff7de9d9-metrics-certs\") pod \"speaker-9sc2x\" (UID: \"c036c09c-bcdf-4fcd-908a-4f59ff7de9d9\") " pod="metallb-system/speaker-9sc2x" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.142151 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c036c09c-bcdf-4fcd-908a-4f59ff7de9d9-metallb-excludel2\") pod \"speaker-9sc2x\" (UID: \"c036c09c-bcdf-4fcd-908a-4f59ff7de9d9\") " pod="metallb-system/speaker-9sc2x" Oct 03 13:07:01 crc kubenswrapper[4578]: E1003 13:07:01.142204 4578 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 03 13:07:01 crc kubenswrapper[4578]: E1003 13:07:01.142274 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c036c09c-bcdf-4fcd-908a-4f59ff7de9d9-memberlist podName:c036c09c-bcdf-4fcd-908a-4f59ff7de9d9 nodeName:}" failed. No retries permitted until 2025-10-03 13:07:01.642257293 +0000 UTC m=+957.440729477 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/c036c09c-bcdf-4fcd-908a-4f59ff7de9d9-memberlist") pod "speaker-9sc2x" (UID: "c036c09c-bcdf-4fcd-908a-4f59ff7de9d9") : secret "metallb-memberlist" not found Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.142593 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-229x9\" (UniqueName: \"kubernetes.io/projected/28ab5270-7890-40e5-8414-fd4a1b9d2742-kube-api-access-229x9\") pod \"controller-68d546b9d8-jhpgk\" (UID: \"28ab5270-7890-40e5-8414-fd4a1b9d2742\") " pod="metallb-system/controller-68d546b9d8-jhpgk" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.142641 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/28ab5270-7890-40e5-8414-fd4a1b9d2742-metrics-certs\") pod \"controller-68d546b9d8-jhpgk\" (UID: \"28ab5270-7890-40e5-8414-fd4a1b9d2742\") " pod="metallb-system/controller-68d546b9d8-jhpgk" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.142667 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28ab5270-7890-40e5-8414-fd4a1b9d2742-cert\") pod \"controller-68d546b9d8-jhpgk\" (UID: \"28ab5270-7890-40e5-8414-fd4a1b9d2742\") " pod="metallb-system/controller-68d546b9d8-jhpgk" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.143081 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/c036c09c-bcdf-4fcd-908a-4f59ff7de9d9-metallb-excludel2\") pod \"speaker-9sc2x\" (UID: \"c036c09c-bcdf-4fcd-908a-4f59ff7de9d9\") " pod="metallb-system/speaker-9sc2x" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.156289 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c036c09c-bcdf-4fcd-908a-4f59ff7de9d9-metrics-certs\") pod \"speaker-9sc2x\" (UID: \"c036c09c-bcdf-4fcd-908a-4f59ff7de9d9\") " pod="metallb-system/speaker-9sc2x" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.158528 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jpxf\" (UniqueName: \"kubernetes.io/projected/c036c09c-bcdf-4fcd-908a-4f59ff7de9d9-kube-api-access-4jpxf\") pod \"speaker-9sc2x\" (UID: \"c036c09c-bcdf-4fcd-908a-4f59ff7de9d9\") " pod="metallb-system/speaker-9sc2x" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.170475 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.247257 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-229x9\" (UniqueName: \"kubernetes.io/projected/28ab5270-7890-40e5-8414-fd4a1b9d2742-kube-api-access-229x9\") pod \"controller-68d546b9d8-jhpgk\" (UID: \"28ab5270-7890-40e5-8414-fd4a1b9d2742\") " pod="metallb-system/controller-68d546b9d8-jhpgk" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.247336 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/28ab5270-7890-40e5-8414-fd4a1b9d2742-metrics-certs\") pod \"controller-68d546b9d8-jhpgk\" (UID: \"28ab5270-7890-40e5-8414-fd4a1b9d2742\") " pod="metallb-system/controller-68d546b9d8-jhpgk" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.247365 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28ab5270-7890-40e5-8414-fd4a1b9d2742-cert\") pod \"controller-68d546b9d8-jhpgk\" (UID: \"28ab5270-7890-40e5-8414-fd4a1b9d2742\") " pod="metallb-system/controller-68d546b9d8-jhpgk" Oct 03 13:07:01 crc kubenswrapper[4578]: E1003 13:07:01.247932 4578 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Oct 03 13:07:01 crc kubenswrapper[4578]: E1003 13:07:01.248119 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/28ab5270-7890-40e5-8414-fd4a1b9d2742-metrics-certs podName:28ab5270-7890-40e5-8414-fd4a1b9d2742 nodeName:}" failed. No retries permitted until 2025-10-03 13:07:01.748080421 +0000 UTC m=+957.546552615 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/28ab5270-7890-40e5-8414-fd4a1b9d2742-metrics-certs") pod "controller-68d546b9d8-jhpgk" (UID: "28ab5270-7890-40e5-8414-fd4a1b9d2742") : secret "controller-certs-secret" not found Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.252091 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28ab5270-7890-40e5-8414-fd4a1b9d2742-cert\") pod \"controller-68d546b9d8-jhpgk\" (UID: \"28ab5270-7890-40e5-8414-fd4a1b9d2742\") " pod="metallb-system/controller-68d546b9d8-jhpgk" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.316832 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-229x9\" (UniqueName: \"kubernetes.io/projected/28ab5270-7890-40e5-8414-fd4a1b9d2742-kube-api-access-229x9\") pod \"controller-68d546b9d8-jhpgk\" (UID: \"28ab5270-7890-40e5-8414-fd4a1b9d2742\") " pod="metallb-system/controller-68d546b9d8-jhpgk" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.551403 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3d7fbaf2-ba49-481f-a56c-df68ef840423-cert\") pod \"frr-k8s-webhook-server-64bf5d555-t7dz4\" (UID: \"3d7fbaf2-ba49-481f-a56c-df68ef840423\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t7dz4" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.555193 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3d7fbaf2-ba49-481f-a56c-df68ef840423-cert\") pod \"frr-k8s-webhook-server-64bf5d555-t7dz4\" (UID: \"3d7fbaf2-ba49-481f-a56c-df68ef840423\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t7dz4" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.652316 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c036c09c-bcdf-4fcd-908a-4f59ff7de9d9-memberlist\") pod \"speaker-9sc2x\" (UID: \"c036c09c-bcdf-4fcd-908a-4f59ff7de9d9\") " pod="metallb-system/speaker-9sc2x" Oct 03 13:07:01 crc kubenswrapper[4578]: E1003 13:07:01.652570 4578 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 03 13:07:01 crc kubenswrapper[4578]: E1003 13:07:01.652622 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c036c09c-bcdf-4fcd-908a-4f59ff7de9d9-memberlist podName:c036c09c-bcdf-4fcd-908a-4f59ff7de9d9 nodeName:}" failed. No retries permitted until 2025-10-03 13:07:02.652608351 +0000 UTC m=+958.451080535 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/c036c09c-bcdf-4fcd-908a-4f59ff7de9d9-memberlist") pod "speaker-9sc2x" (UID: "c036c09c-bcdf-4fcd-908a-4f59ff7de9d9") : secret "metallb-memberlist" not found Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.754201 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/28ab5270-7890-40e5-8414-fd4a1b9d2742-metrics-certs\") pod \"controller-68d546b9d8-jhpgk\" (UID: \"28ab5270-7890-40e5-8414-fd4a1b9d2742\") " pod="metallb-system/controller-68d546b9d8-jhpgk" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.757505 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/28ab5270-7890-40e5-8414-fd4a1b9d2742-metrics-certs\") pod \"controller-68d546b9d8-jhpgk\" (UID: \"28ab5270-7890-40e5-8414-fd4a1b9d2742\") " pod="metallb-system/controller-68d546b9d8-jhpgk" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.783366 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t7dz4" Oct 03 13:07:01 crc kubenswrapper[4578]: I1003 13:07:01.898223 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-jhpgk" Oct 03 13:07:02 crc kubenswrapper[4578]: I1003 13:07:02.104561 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-t7dz4"] Oct 03 13:07:02 crc kubenswrapper[4578]: I1003 13:07:02.155110 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-jhpgk"] Oct 03 13:07:02 crc kubenswrapper[4578]: I1003 13:07:02.647163 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vcftq" event={"ID":"70df60c7-212d-4834-b237-527ca5710600","Type":"ContainerStarted","Data":"67ecc006b7746fc122242fa107d293e446625aa72fca04612542cbcfd974c891"} Oct 03 13:07:02 crc kubenswrapper[4578]: I1003 13:07:02.648066 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t7dz4" event={"ID":"3d7fbaf2-ba49-481f-a56c-df68ef840423","Type":"ContainerStarted","Data":"5642096f426c81f3c9baabe2abb4e45babcc6f7485b497a7611797e0031d88bd"} Oct 03 13:07:02 crc kubenswrapper[4578]: I1003 13:07:02.650878 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-jhpgk" event={"ID":"28ab5270-7890-40e5-8414-fd4a1b9d2742","Type":"ContainerStarted","Data":"5b08f3ddaac028bbd36966da2af02ecf49ff4beb905dc08bf3f13ac4f1b21a04"} Oct 03 13:07:02 crc kubenswrapper[4578]: I1003 13:07:02.650906 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-jhpgk" event={"ID":"28ab5270-7890-40e5-8414-fd4a1b9d2742","Type":"ContainerStarted","Data":"ca4305e95321e05d840e41341004f958d53721300cd56bb2e03b97577685ef7c"} Oct 03 13:07:02 crc kubenswrapper[4578]: I1003 13:07:02.650925 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-jhpgk" event={"ID":"28ab5270-7890-40e5-8414-fd4a1b9d2742","Type":"ContainerStarted","Data":"b8b73a12c523385da6e9319c55e6282c5dcbad334e11ed449dbc6a91ea036100"} Oct 03 13:07:02 crc kubenswrapper[4578]: I1003 13:07:02.651030 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-jhpgk" Oct 03 13:07:02 crc kubenswrapper[4578]: I1003 13:07:02.667342 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c036c09c-bcdf-4fcd-908a-4f59ff7de9d9-memberlist\") pod \"speaker-9sc2x\" (UID: \"c036c09c-bcdf-4fcd-908a-4f59ff7de9d9\") " pod="metallb-system/speaker-9sc2x" Oct 03 13:07:02 crc kubenswrapper[4578]: I1003 13:07:02.670746 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-jhpgk" podStartSLOduration=2.670726315 podStartE2EDuration="2.670726315s" podCreationTimestamp="2025-10-03 13:07:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:07:02.667013438 +0000 UTC m=+958.465485622" watchObservedRunningTime="2025-10-03 13:07:02.670726315 +0000 UTC m=+958.469198499" Oct 03 13:07:02 crc kubenswrapper[4578]: I1003 13:07:02.679692 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/c036c09c-bcdf-4fcd-908a-4f59ff7de9d9-memberlist\") pod \"speaker-9sc2x\" (UID: \"c036c09c-bcdf-4fcd-908a-4f59ff7de9d9\") " pod="metallb-system/speaker-9sc2x" Oct 03 13:07:02 crc kubenswrapper[4578]: I1003 13:07:02.783407 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-9sc2x" Oct 03 13:07:02 crc kubenswrapper[4578]: W1003 13:07:02.805388 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc036c09c_bcdf_4fcd_908a_4f59ff7de9d9.slice/crio-b64a992f850e9716ae40edfc7bbc3a72987360e76138b4eb88f767f7e962858f WatchSource:0}: Error finding container b64a992f850e9716ae40edfc7bbc3a72987360e76138b4eb88f767f7e962858f: Status 404 returned error can't find the container with id b64a992f850e9716ae40edfc7bbc3a72987360e76138b4eb88f767f7e962858f Oct 03 13:07:03 crc kubenswrapper[4578]: I1003 13:07:03.669818 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9sc2x" event={"ID":"c036c09c-bcdf-4fcd-908a-4f59ff7de9d9","Type":"ContainerStarted","Data":"71e4824b965d123bf14096fe4d23e569d9b036560c02063afce21f147dfe8e3a"} Oct 03 13:07:03 crc kubenswrapper[4578]: I1003 13:07:03.670150 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9sc2x" event={"ID":"c036c09c-bcdf-4fcd-908a-4f59ff7de9d9","Type":"ContainerStarted","Data":"dc977513d8f59f0630976e91b1ac0ee8202eb4884fd811f3886410c4dc0620a5"} Oct 03 13:07:03 crc kubenswrapper[4578]: I1003 13:07:03.670162 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9sc2x" event={"ID":"c036c09c-bcdf-4fcd-908a-4f59ff7de9d9","Type":"ContainerStarted","Data":"b64a992f850e9716ae40edfc7bbc3a72987360e76138b4eb88f767f7e962858f"} Oct 03 13:07:03 crc kubenswrapper[4578]: I1003 13:07:03.670813 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-9sc2x" Oct 03 13:07:03 crc kubenswrapper[4578]: I1003 13:07:03.693897 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-9sc2x" podStartSLOduration=3.693883988 podStartE2EDuration="3.693883988s" podCreationTimestamp="2025-10-03 13:07:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:07:03.692244066 +0000 UTC m=+959.490716250" watchObservedRunningTime="2025-10-03 13:07:03.693883988 +0000 UTC m=+959.492356172" Oct 03 13:07:10 crc kubenswrapper[4578]: I1003 13:07:10.712040 4578 generic.go:334] "Generic (PLEG): container finished" podID="70df60c7-212d-4834-b237-527ca5710600" containerID="fc98e48bc63ef98ea298ba0ee98147b18b0dfe65b43deb8363abd6cdfeb878e4" exitCode=0 Oct 03 13:07:10 crc kubenswrapper[4578]: I1003 13:07:10.712129 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vcftq" event={"ID":"70df60c7-212d-4834-b237-527ca5710600","Type":"ContainerDied","Data":"fc98e48bc63ef98ea298ba0ee98147b18b0dfe65b43deb8363abd6cdfeb878e4"} Oct 03 13:07:11 crc kubenswrapper[4578]: I1003 13:07:11.722067 4578 generic.go:334] "Generic (PLEG): container finished" podID="70df60c7-212d-4834-b237-527ca5710600" containerID="1794651d705c945e11cdcc57b8c4c9c72dd2a5d1956225cf32e21b05159a35f2" exitCode=0 Oct 03 13:07:11 crc kubenswrapper[4578]: I1003 13:07:11.722119 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vcftq" event={"ID":"70df60c7-212d-4834-b237-527ca5710600","Type":"ContainerDied","Data":"1794651d705c945e11cdcc57b8c4c9c72dd2a5d1956225cf32e21b05159a35f2"} Oct 03 13:07:12 crc kubenswrapper[4578]: I1003 13:07:12.730841 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vcftq" event={"ID":"70df60c7-212d-4834-b237-527ca5710600","Type":"ContainerDied","Data":"7b99b10224122ee33ed98db7b2cd2573f142ab67884e518a8752e84ee502a9f0"} Oct 03 13:07:12 crc kubenswrapper[4578]: I1003 13:07:12.730794 4578 generic.go:334] "Generic (PLEG): container finished" podID="70df60c7-212d-4834-b237-527ca5710600" containerID="7b99b10224122ee33ed98db7b2cd2573f142ab67884e518a8752e84ee502a9f0" exitCode=0 Oct 03 13:07:12 crc kubenswrapper[4578]: I1003 13:07:12.732731 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t7dz4" event={"ID":"3d7fbaf2-ba49-481f-a56c-df68ef840423","Type":"ContainerStarted","Data":"a80ca3443bd135b158cedde084d14a8d8571cd0fe1b89e9ef25a23440310fdc9"} Oct 03 13:07:12 crc kubenswrapper[4578]: I1003 13:07:12.732877 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t7dz4" Oct 03 13:07:12 crc kubenswrapper[4578]: I1003 13:07:12.774109 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t7dz4" podStartSLOduration=2.803164033 podStartE2EDuration="12.774090671s" podCreationTimestamp="2025-10-03 13:07:00 +0000 UTC" firstStartedPulling="2025-10-03 13:07:02.108688747 +0000 UTC m=+957.907160931" lastFinishedPulling="2025-10-03 13:07:12.079615365 +0000 UTC m=+967.878087569" observedRunningTime="2025-10-03 13:07:12.771998075 +0000 UTC m=+968.570470259" watchObservedRunningTime="2025-10-03 13:07:12.774090671 +0000 UTC m=+968.572562865" Oct 03 13:07:13 crc kubenswrapper[4578]: I1003 13:07:13.741470 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vcftq" event={"ID":"70df60c7-212d-4834-b237-527ca5710600","Type":"ContainerStarted","Data":"c0c95a89c90d7a84e5fc16e5e4a1d2549d7b0f48c7b7f5310ac677ea1fb7ee23"} Oct 03 13:07:13 crc kubenswrapper[4578]: I1003 13:07:13.741813 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vcftq" event={"ID":"70df60c7-212d-4834-b237-527ca5710600","Type":"ContainerStarted","Data":"f07adfd89a62ce4be951da75f75a94d91f5c15e3fb066950b968ebc3e9fb6f09"} Oct 03 13:07:13 crc kubenswrapper[4578]: I1003 13:07:13.741823 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vcftq" event={"ID":"70df60c7-212d-4834-b237-527ca5710600","Type":"ContainerStarted","Data":"46bca5e56b812254c324f2243bd5aee2bb8edf590216840348b9bad597d02e1e"} Oct 03 13:07:13 crc kubenswrapper[4578]: I1003 13:07:13.741832 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vcftq" event={"ID":"70df60c7-212d-4834-b237-527ca5710600","Type":"ContainerStarted","Data":"2eeeb07ad2bf4c8ae684491b02edd811825d68867e9a2ca49bc9008836f7fcec"} Oct 03 13:07:13 crc kubenswrapper[4578]: I1003 13:07:13.741840 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vcftq" event={"ID":"70df60c7-212d-4834-b237-527ca5710600","Type":"ContainerStarted","Data":"2cf3a3ef75c8b374f73ba6e2570d58925dde331b1ebc2c247bb4ca77deb8a534"} Oct 03 13:07:13 crc kubenswrapper[4578]: I1003 13:07:13.741849 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vcftq" event={"ID":"70df60c7-212d-4834-b237-527ca5710600","Type":"ContainerStarted","Data":"ef7ae7a47fdecd15e02b5786890508bdd54f37bf62f5335cd2d06d214911ad9c"} Oct 03 13:07:13 crc kubenswrapper[4578]: I1003 13:07:13.766422 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-vcftq" podStartSLOduration=5.899072985 podStartE2EDuration="13.766403481s" podCreationTimestamp="2025-10-03 13:07:00 +0000 UTC" firstStartedPulling="2025-10-03 13:07:01.733480502 +0000 UTC m=+957.531952686" lastFinishedPulling="2025-10-03 13:07:09.600810998 +0000 UTC m=+965.399283182" observedRunningTime="2025-10-03 13:07:13.765493713 +0000 UTC m=+969.563965897" watchObservedRunningTime="2025-10-03 13:07:13.766403481 +0000 UTC m=+969.564875655" Oct 03 13:07:14 crc kubenswrapper[4578]: I1003 13:07:14.746735 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:16 crc kubenswrapper[4578]: I1003 13:07:16.172061 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:16 crc kubenswrapper[4578]: I1003 13:07:16.208030 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:21 crc kubenswrapper[4578]: I1003 13:07:21.902137 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-jhpgk" Oct 03 13:07:22 crc kubenswrapper[4578]: I1003 13:07:22.788183 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-9sc2x" Oct 03 13:07:25 crc kubenswrapper[4578]: I1003 13:07:25.782923 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-qpcpl"] Oct 03 13:07:25 crc kubenswrapper[4578]: I1003 13:07:25.784004 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qpcpl" Oct 03 13:07:25 crc kubenswrapper[4578]: I1003 13:07:25.795852 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 03 13:07:25 crc kubenswrapper[4578]: I1003 13:07:25.795941 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-r2cpk" Oct 03 13:07:25 crc kubenswrapper[4578]: I1003 13:07:25.796097 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 03 13:07:25 crc kubenswrapper[4578]: I1003 13:07:25.819416 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qpcpl"] Oct 03 13:07:25 crc kubenswrapper[4578]: I1003 13:07:25.871087 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78q26\" (UniqueName: \"kubernetes.io/projected/0eb25365-4414-400f-be30-d5958d0651c0-kube-api-access-78q26\") pod \"openstack-operator-index-qpcpl\" (UID: \"0eb25365-4414-400f-be30-d5958d0651c0\") " pod="openstack-operators/openstack-operator-index-qpcpl" Oct 03 13:07:25 crc kubenswrapper[4578]: I1003 13:07:25.972313 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78q26\" (UniqueName: \"kubernetes.io/projected/0eb25365-4414-400f-be30-d5958d0651c0-kube-api-access-78q26\") pod \"openstack-operator-index-qpcpl\" (UID: \"0eb25365-4414-400f-be30-d5958d0651c0\") " pod="openstack-operators/openstack-operator-index-qpcpl" Oct 03 13:07:25 crc kubenswrapper[4578]: I1003 13:07:25.992469 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78q26\" (UniqueName: \"kubernetes.io/projected/0eb25365-4414-400f-be30-d5958d0651c0-kube-api-access-78q26\") pod \"openstack-operator-index-qpcpl\" (UID: \"0eb25365-4414-400f-be30-d5958d0651c0\") " pod="openstack-operators/openstack-operator-index-qpcpl" Oct 03 13:07:26 crc kubenswrapper[4578]: I1003 13:07:26.105444 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qpcpl" Oct 03 13:07:26 crc kubenswrapper[4578]: I1003 13:07:26.488311 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qpcpl"] Oct 03 13:07:26 crc kubenswrapper[4578]: W1003 13:07:26.497330 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0eb25365_4414_400f_be30_d5958d0651c0.slice/crio-f3ff49cb9833b0130a90ad89b836e33c14e8b489be3681c2231d6b01cd6e17b7 WatchSource:0}: Error finding container f3ff49cb9833b0130a90ad89b836e33c14e8b489be3681c2231d6b01cd6e17b7: Status 404 returned error can't find the container with id f3ff49cb9833b0130a90ad89b836e33c14e8b489be3681c2231d6b01cd6e17b7 Oct 03 13:07:26 crc kubenswrapper[4578]: I1003 13:07:26.821178 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qpcpl" event={"ID":"0eb25365-4414-400f-be30-d5958d0651c0","Type":"ContainerStarted","Data":"f3ff49cb9833b0130a90ad89b836e33c14e8b489be3681c2231d6b01cd6e17b7"} Oct 03 13:07:27 crc kubenswrapper[4578]: I1003 13:07:27.739167 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-qpcpl"] Oct 03 13:07:28 crc kubenswrapper[4578]: I1003 13:07:28.141209 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-svcqt"] Oct 03 13:07:28 crc kubenswrapper[4578]: I1003 13:07:28.144049 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-svcqt" Oct 03 13:07:28 crc kubenswrapper[4578]: I1003 13:07:28.171020 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-svcqt"] Oct 03 13:07:28 crc kubenswrapper[4578]: I1003 13:07:28.206780 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pscq4\" (UniqueName: \"kubernetes.io/projected/49c934f0-4e15-40cf-a0e6-f067f39912f7-kube-api-access-pscq4\") pod \"openstack-operator-index-svcqt\" (UID: \"49c934f0-4e15-40cf-a0e6-f067f39912f7\") " pod="openstack-operators/openstack-operator-index-svcqt" Oct 03 13:07:28 crc kubenswrapper[4578]: I1003 13:07:28.307510 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pscq4\" (UniqueName: \"kubernetes.io/projected/49c934f0-4e15-40cf-a0e6-f067f39912f7-kube-api-access-pscq4\") pod \"openstack-operator-index-svcqt\" (UID: \"49c934f0-4e15-40cf-a0e6-f067f39912f7\") " pod="openstack-operators/openstack-operator-index-svcqt" Oct 03 13:07:28 crc kubenswrapper[4578]: I1003 13:07:28.340647 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pscq4\" (UniqueName: \"kubernetes.io/projected/49c934f0-4e15-40cf-a0e6-f067f39912f7-kube-api-access-pscq4\") pod \"openstack-operator-index-svcqt\" (UID: \"49c934f0-4e15-40cf-a0e6-f067f39912f7\") " pod="openstack-operators/openstack-operator-index-svcqt" Oct 03 13:07:28 crc kubenswrapper[4578]: I1003 13:07:28.474396 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-svcqt" Oct 03 13:07:29 crc kubenswrapper[4578]: I1003 13:07:29.904564 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-svcqt"] Oct 03 13:07:30 crc kubenswrapper[4578]: I1003 13:07:30.842614 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-svcqt" event={"ID":"49c934f0-4e15-40cf-a0e6-f067f39912f7","Type":"ContainerStarted","Data":"4f51a977aa7a5a87e1f95b355ecea8ee23b6c73096ac8ee91c685aa860bef8c2"} Oct 03 13:07:30 crc kubenswrapper[4578]: I1003 13:07:30.843135 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-svcqt" event={"ID":"49c934f0-4e15-40cf-a0e6-f067f39912f7","Type":"ContainerStarted","Data":"54baf53d31dd9df3ec9766e72c190a57a6abb9665d688424507ffa7e3fc8a5a2"} Oct 03 13:07:30 crc kubenswrapper[4578]: I1003 13:07:30.843681 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qpcpl" event={"ID":"0eb25365-4414-400f-be30-d5958d0651c0","Type":"ContainerStarted","Data":"93a9077d68ad7be0c9be75a3b765565c71877ebf3fec23c4e31f9407951846d9"} Oct 03 13:07:30 crc kubenswrapper[4578]: I1003 13:07:30.843724 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-qpcpl" podUID="0eb25365-4414-400f-be30-d5958d0651c0" containerName="registry-server" containerID="cri-o://93a9077d68ad7be0c9be75a3b765565c71877ebf3fec23c4e31f9407951846d9" gracePeriod=2 Oct 03 13:07:30 crc kubenswrapper[4578]: I1003 13:07:30.863871 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-svcqt" podStartSLOduration=2.803515155 podStartE2EDuration="2.863847528s" podCreationTimestamp="2025-10-03 13:07:28 +0000 UTC" firstStartedPulling="2025-10-03 13:07:30.086083635 +0000 UTC m=+985.884555819" lastFinishedPulling="2025-10-03 13:07:30.146416008 +0000 UTC m=+985.944888192" observedRunningTime="2025-10-03 13:07:30.860462491 +0000 UTC m=+986.658934675" watchObservedRunningTime="2025-10-03 13:07:30.863847528 +0000 UTC m=+986.662319722" Oct 03 13:07:30 crc kubenswrapper[4578]: I1003 13:07:30.883542 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-qpcpl" podStartSLOduration=2.28935496 podStartE2EDuration="5.883521909s" podCreationTimestamp="2025-10-03 13:07:25 +0000 UTC" firstStartedPulling="2025-10-03 13:07:26.499508766 +0000 UTC m=+982.297980950" lastFinishedPulling="2025-10-03 13:07:30.093675715 +0000 UTC m=+985.892147899" observedRunningTime="2025-10-03 13:07:30.879001306 +0000 UTC m=+986.677473510" watchObservedRunningTime="2025-10-03 13:07:30.883521909 +0000 UTC m=+986.681994093" Oct 03 13:07:31 crc kubenswrapper[4578]: I1003 13:07:31.174015 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-vcftq" Oct 03 13:07:31 crc kubenswrapper[4578]: I1003 13:07:31.186288 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qpcpl" Oct 03 13:07:31 crc kubenswrapper[4578]: I1003 13:07:31.244266 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78q26\" (UniqueName: \"kubernetes.io/projected/0eb25365-4414-400f-be30-d5958d0651c0-kube-api-access-78q26\") pod \"0eb25365-4414-400f-be30-d5958d0651c0\" (UID: \"0eb25365-4414-400f-be30-d5958d0651c0\") " Oct 03 13:07:31 crc kubenswrapper[4578]: I1003 13:07:31.259592 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0eb25365-4414-400f-be30-d5958d0651c0-kube-api-access-78q26" (OuterVolumeSpecName: "kube-api-access-78q26") pod "0eb25365-4414-400f-be30-d5958d0651c0" (UID: "0eb25365-4414-400f-be30-d5958d0651c0"). InnerVolumeSpecName "kube-api-access-78q26". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:07:31 crc kubenswrapper[4578]: I1003 13:07:31.346731 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78q26\" (UniqueName: \"kubernetes.io/projected/0eb25365-4414-400f-be30-d5958d0651c0-kube-api-access-78q26\") on node \"crc\" DevicePath \"\"" Oct 03 13:07:31 crc kubenswrapper[4578]: I1003 13:07:31.786692 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t7dz4" Oct 03 13:07:31 crc kubenswrapper[4578]: I1003 13:07:31.854212 4578 generic.go:334] "Generic (PLEG): container finished" podID="0eb25365-4414-400f-be30-d5958d0651c0" containerID="93a9077d68ad7be0c9be75a3b765565c71877ebf3fec23c4e31f9407951846d9" exitCode=0 Oct 03 13:07:31 crc kubenswrapper[4578]: I1003 13:07:31.854709 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qpcpl" event={"ID":"0eb25365-4414-400f-be30-d5958d0651c0","Type":"ContainerDied","Data":"93a9077d68ad7be0c9be75a3b765565c71877ebf3fec23c4e31f9407951846d9"} Oct 03 13:07:31 crc kubenswrapper[4578]: I1003 13:07:31.854755 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qpcpl" Oct 03 13:07:31 crc kubenswrapper[4578]: I1003 13:07:31.854787 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qpcpl" event={"ID":"0eb25365-4414-400f-be30-d5958d0651c0","Type":"ContainerDied","Data":"f3ff49cb9833b0130a90ad89b836e33c14e8b489be3681c2231d6b01cd6e17b7"} Oct 03 13:07:31 crc kubenswrapper[4578]: I1003 13:07:31.854809 4578 scope.go:117] "RemoveContainer" containerID="93a9077d68ad7be0c9be75a3b765565c71877ebf3fec23c4e31f9407951846d9" Oct 03 13:07:31 crc kubenswrapper[4578]: I1003 13:07:31.872717 4578 scope.go:117] "RemoveContainer" containerID="93a9077d68ad7be0c9be75a3b765565c71877ebf3fec23c4e31f9407951846d9" Oct 03 13:07:31 crc kubenswrapper[4578]: E1003 13:07:31.873046 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93a9077d68ad7be0c9be75a3b765565c71877ebf3fec23c4e31f9407951846d9\": container with ID starting with 93a9077d68ad7be0c9be75a3b765565c71877ebf3fec23c4e31f9407951846d9 not found: ID does not exist" containerID="93a9077d68ad7be0c9be75a3b765565c71877ebf3fec23c4e31f9407951846d9" Oct 03 13:07:31 crc kubenswrapper[4578]: I1003 13:07:31.873070 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93a9077d68ad7be0c9be75a3b765565c71877ebf3fec23c4e31f9407951846d9"} err="failed to get container status \"93a9077d68ad7be0c9be75a3b765565c71877ebf3fec23c4e31f9407951846d9\": rpc error: code = NotFound desc = could not find container \"93a9077d68ad7be0c9be75a3b765565c71877ebf3fec23c4e31f9407951846d9\": container with ID starting with 93a9077d68ad7be0c9be75a3b765565c71877ebf3fec23c4e31f9407951846d9 not found: ID does not exist" Oct 03 13:07:31 crc kubenswrapper[4578]: I1003 13:07:31.885047 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-qpcpl"] Oct 03 13:07:31 crc kubenswrapper[4578]: I1003 13:07:31.888163 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-qpcpl"] Oct 03 13:07:32 crc kubenswrapper[4578]: I1003 13:07:32.918041 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0eb25365-4414-400f-be30-d5958d0651c0" path="/var/lib/kubelet/pods/0eb25365-4414-400f-be30-d5958d0651c0/volumes" Oct 03 13:07:38 crc kubenswrapper[4578]: I1003 13:07:38.475015 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-svcqt" Oct 03 13:07:38 crc kubenswrapper[4578]: I1003 13:07:38.475329 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-svcqt" Oct 03 13:07:38 crc kubenswrapper[4578]: I1003 13:07:38.501514 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-svcqt" Oct 03 13:07:38 crc kubenswrapper[4578]: I1003 13:07:38.918192 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-svcqt" Oct 03 13:07:46 crc kubenswrapper[4578]: I1003 13:07:46.649747 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9"] Oct 03 13:07:46 crc kubenswrapper[4578]: E1003 13:07:46.650592 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0eb25365-4414-400f-be30-d5958d0651c0" containerName="registry-server" Oct 03 13:07:46 crc kubenswrapper[4578]: I1003 13:07:46.650609 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="0eb25365-4414-400f-be30-d5958d0651c0" containerName="registry-server" Oct 03 13:07:46 crc kubenswrapper[4578]: I1003 13:07:46.650759 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="0eb25365-4414-400f-be30-d5958d0651c0" containerName="registry-server" Oct 03 13:07:46 crc kubenswrapper[4578]: I1003 13:07:46.651532 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9" Oct 03 13:07:46 crc kubenswrapper[4578]: I1003 13:07:46.654035 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-swgpv" Oct 03 13:07:46 crc kubenswrapper[4578]: I1003 13:07:46.661807 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9"] Oct 03 13:07:46 crc kubenswrapper[4578]: I1003 13:07:46.746598 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d-bundle\") pod \"6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9\" (UID: \"cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d\") " pod="openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9" Oct 03 13:07:46 crc kubenswrapper[4578]: I1003 13:07:46.747028 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzvws\" (UniqueName: \"kubernetes.io/projected/cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d-kube-api-access-qzvws\") pod \"6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9\" (UID: \"cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d\") " pod="openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9" Oct 03 13:07:46 crc kubenswrapper[4578]: I1003 13:07:46.747211 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d-util\") pod \"6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9\" (UID: \"cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d\") " pod="openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9" Oct 03 13:07:46 crc kubenswrapper[4578]: I1003 13:07:46.848333 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d-bundle\") pod \"6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9\" (UID: \"cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d\") " pod="openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9" Oct 03 13:07:46 crc kubenswrapper[4578]: I1003 13:07:46.848412 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzvws\" (UniqueName: \"kubernetes.io/projected/cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d-kube-api-access-qzvws\") pod \"6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9\" (UID: \"cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d\") " pod="openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9" Oct 03 13:07:46 crc kubenswrapper[4578]: I1003 13:07:46.848434 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d-util\") pod \"6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9\" (UID: \"cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d\") " pod="openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9" Oct 03 13:07:46 crc kubenswrapper[4578]: I1003 13:07:46.848930 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d-bundle\") pod \"6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9\" (UID: \"cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d\") " pod="openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9" Oct 03 13:07:46 crc kubenswrapper[4578]: I1003 13:07:46.848976 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d-util\") pod \"6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9\" (UID: \"cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d\") " pod="openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9" Oct 03 13:07:46 crc kubenswrapper[4578]: I1003 13:07:46.869312 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzvws\" (UniqueName: \"kubernetes.io/projected/cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d-kube-api-access-qzvws\") pod \"6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9\" (UID: \"cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d\") " pod="openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9" Oct 03 13:07:46 crc kubenswrapper[4578]: I1003 13:07:46.966151 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9" Oct 03 13:07:47 crc kubenswrapper[4578]: I1003 13:07:47.356824 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9"] Oct 03 13:07:47 crc kubenswrapper[4578]: W1003 13:07:47.365193 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd645ad4_0d57_4dc2_9b0b_a6e025a5be5d.slice/crio-249bd1e4253dff8a4b765714868f8dda8ea0e2b07a818e73b32c735d3e52cd21 WatchSource:0}: Error finding container 249bd1e4253dff8a4b765714868f8dda8ea0e2b07a818e73b32c735d3e52cd21: Status 404 returned error can't find the container with id 249bd1e4253dff8a4b765714868f8dda8ea0e2b07a818e73b32c735d3e52cd21 Oct 03 13:07:47 crc kubenswrapper[4578]: I1003 13:07:47.946223 4578 generic.go:334] "Generic (PLEG): container finished" podID="cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d" containerID="1b68b06e0ae276544b4bb612ba09e7306fd3d5001c3249e148282b412fb15c05" exitCode=0 Oct 03 13:07:47 crc kubenswrapper[4578]: I1003 13:07:47.946287 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9" event={"ID":"cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d","Type":"ContainerDied","Data":"1b68b06e0ae276544b4bb612ba09e7306fd3d5001c3249e148282b412fb15c05"} Oct 03 13:07:47 crc kubenswrapper[4578]: I1003 13:07:47.946534 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9" event={"ID":"cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d","Type":"ContainerStarted","Data":"249bd1e4253dff8a4b765714868f8dda8ea0e2b07a818e73b32c735d3e52cd21"} Oct 03 13:07:48 crc kubenswrapper[4578]: I1003 13:07:48.954615 4578 generic.go:334] "Generic (PLEG): container finished" podID="cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d" containerID="facedf8c5477a340f741307c162896bc62e424cb8e7775a0f605d4cd7e03671f" exitCode=0 Oct 03 13:07:48 crc kubenswrapper[4578]: I1003 13:07:48.954724 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9" event={"ID":"cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d","Type":"ContainerDied","Data":"facedf8c5477a340f741307c162896bc62e424cb8e7775a0f605d4cd7e03671f"} Oct 03 13:07:49 crc kubenswrapper[4578]: I1003 13:07:49.962932 4578 generic.go:334] "Generic (PLEG): container finished" podID="cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d" containerID="b94b437c6ed80aba10f1fd82ae0c1b7d7be0bcd62f856b8b7c4eed57404d6477" exitCode=0 Oct 03 13:07:49 crc kubenswrapper[4578]: I1003 13:07:49.962978 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9" event={"ID":"cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d","Type":"ContainerDied","Data":"b94b437c6ed80aba10f1fd82ae0c1b7d7be0bcd62f856b8b7c4eed57404d6477"} Oct 03 13:07:51 crc kubenswrapper[4578]: I1003 13:07:51.248937 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9" Oct 03 13:07:51 crc kubenswrapper[4578]: I1003 13:07:51.304811 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzvws\" (UniqueName: \"kubernetes.io/projected/cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d-kube-api-access-qzvws\") pod \"cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d\" (UID: \"cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d\") " Oct 03 13:07:51 crc kubenswrapper[4578]: I1003 13:07:51.304971 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d-bundle\") pod \"cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d\" (UID: \"cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d\") " Oct 03 13:07:51 crc kubenswrapper[4578]: I1003 13:07:51.305036 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d-util\") pod \"cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d\" (UID: \"cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d\") " Oct 03 13:07:51 crc kubenswrapper[4578]: I1003 13:07:51.306017 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d-bundle" (OuterVolumeSpecName: "bundle") pod "cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d" (UID: "cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:07:51 crc kubenswrapper[4578]: I1003 13:07:51.309356 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d-kube-api-access-qzvws" (OuterVolumeSpecName: "kube-api-access-qzvws") pod "cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d" (UID: "cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d"). InnerVolumeSpecName "kube-api-access-qzvws". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:07:51 crc kubenswrapper[4578]: I1003 13:07:51.335197 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d-util" (OuterVolumeSpecName: "util") pod "cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d" (UID: "cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:07:51 crc kubenswrapper[4578]: I1003 13:07:51.406010 4578 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d-util\") on node \"crc\" DevicePath \"\"" Oct 03 13:07:51 crc kubenswrapper[4578]: I1003 13:07:51.406041 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzvws\" (UniqueName: \"kubernetes.io/projected/cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d-kube-api-access-qzvws\") on node \"crc\" DevicePath \"\"" Oct 03 13:07:51 crc kubenswrapper[4578]: I1003 13:07:51.406054 4578 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:07:51 crc kubenswrapper[4578]: I1003 13:07:51.975668 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9" event={"ID":"cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d","Type":"ContainerDied","Data":"249bd1e4253dff8a4b765714868f8dda8ea0e2b07a818e73b32c735d3e52cd21"} Oct 03 13:07:51 crc kubenswrapper[4578]: I1003 13:07:51.975717 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="249bd1e4253dff8a4b765714868f8dda8ea0e2b07a818e73b32c735d3e52cd21" Oct 03 13:07:51 crc kubenswrapper[4578]: I1003 13:07:51.975736 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9" Oct 03 13:07:59 crc kubenswrapper[4578]: I1003 13:07:59.080613 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-764f84468b-x95lq"] Oct 03 13:07:59 crc kubenswrapper[4578]: E1003 13:07:59.081359 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d" containerName="pull" Oct 03 13:07:59 crc kubenswrapper[4578]: I1003 13:07:59.081370 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d" containerName="pull" Oct 03 13:07:59 crc kubenswrapper[4578]: E1003 13:07:59.081388 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d" containerName="extract" Oct 03 13:07:59 crc kubenswrapper[4578]: I1003 13:07:59.081394 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d" containerName="extract" Oct 03 13:07:59 crc kubenswrapper[4578]: E1003 13:07:59.081401 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d" containerName="util" Oct 03 13:07:59 crc kubenswrapper[4578]: I1003 13:07:59.081408 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d" containerName="util" Oct 03 13:07:59 crc kubenswrapper[4578]: I1003 13:07:59.081509 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d" containerName="extract" Oct 03 13:07:59 crc kubenswrapper[4578]: I1003 13:07:59.082117 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-764f84468b-x95lq" Oct 03 13:07:59 crc kubenswrapper[4578]: I1003 13:07:59.083455 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-tc57p" Oct 03 13:07:59 crc kubenswrapper[4578]: I1003 13:07:59.095041 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-764f84468b-x95lq"] Oct 03 13:07:59 crc kubenswrapper[4578]: I1003 13:07:59.206585 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbmbw\" (UniqueName: \"kubernetes.io/projected/440efe1c-5d5c-4e97-a684-3afa7a78b8d7-kube-api-access-qbmbw\") pod \"openstack-operator-controller-operator-764f84468b-x95lq\" (UID: \"440efe1c-5d5c-4e97-a684-3afa7a78b8d7\") " pod="openstack-operators/openstack-operator-controller-operator-764f84468b-x95lq" Oct 03 13:07:59 crc kubenswrapper[4578]: I1003 13:07:59.308146 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbmbw\" (UniqueName: \"kubernetes.io/projected/440efe1c-5d5c-4e97-a684-3afa7a78b8d7-kube-api-access-qbmbw\") pod \"openstack-operator-controller-operator-764f84468b-x95lq\" (UID: \"440efe1c-5d5c-4e97-a684-3afa7a78b8d7\") " pod="openstack-operators/openstack-operator-controller-operator-764f84468b-x95lq" Oct 03 13:07:59 crc kubenswrapper[4578]: I1003 13:07:59.327174 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbmbw\" (UniqueName: \"kubernetes.io/projected/440efe1c-5d5c-4e97-a684-3afa7a78b8d7-kube-api-access-qbmbw\") pod \"openstack-operator-controller-operator-764f84468b-x95lq\" (UID: \"440efe1c-5d5c-4e97-a684-3afa7a78b8d7\") " pod="openstack-operators/openstack-operator-controller-operator-764f84468b-x95lq" Oct 03 13:07:59 crc kubenswrapper[4578]: I1003 13:07:59.397192 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-764f84468b-x95lq" Oct 03 13:07:59 crc kubenswrapper[4578]: I1003 13:07:59.843985 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-764f84468b-x95lq"] Oct 03 13:08:00 crc kubenswrapper[4578]: I1003 13:08:00.019806 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-764f84468b-x95lq" event={"ID":"440efe1c-5d5c-4e97-a684-3afa7a78b8d7","Type":"ContainerStarted","Data":"7588b253bffb88f910009a07cb6f54c485d83d1f8da907ef5bd94a9de6f2f1ea"} Oct 03 13:08:06 crc kubenswrapper[4578]: I1003 13:08:06.072413 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-764f84468b-x95lq" event={"ID":"440efe1c-5d5c-4e97-a684-3afa7a78b8d7","Type":"ContainerStarted","Data":"1a8d60db493c15927e3bd778188e86421e2a8fc7ef0ea1ed691158cdd8b60b39"} Oct 03 13:08:08 crc kubenswrapper[4578]: I1003 13:08:08.087447 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-764f84468b-x95lq" event={"ID":"440efe1c-5d5c-4e97-a684-3afa7a78b8d7","Type":"ContainerStarted","Data":"6d96b2cfc176369cb611fea9aed1833b7a8ed856050842120d1f808959138ab5"} Oct 03 13:08:08 crc kubenswrapper[4578]: I1003 13:08:08.087781 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-764f84468b-x95lq" Oct 03 13:08:08 crc kubenswrapper[4578]: I1003 13:08:08.112075 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-764f84468b-x95lq" podStartSLOduration=1.293131193 podStartE2EDuration="9.112060792s" podCreationTimestamp="2025-10-03 13:07:59 +0000 UTC" firstStartedPulling="2025-10-03 13:07:59.851503662 +0000 UTC m=+1015.649975846" lastFinishedPulling="2025-10-03 13:08:07.670433261 +0000 UTC m=+1023.468905445" observedRunningTime="2025-10-03 13:08:08.110724109 +0000 UTC m=+1023.909196313" watchObservedRunningTime="2025-10-03 13:08:08.112060792 +0000 UTC m=+1023.910532976" Oct 03 13:08:19 crc kubenswrapper[4578]: I1003 13:08:19.399790 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-764f84468b-x95lq" Oct 03 13:08:35 crc kubenswrapper[4578]: I1003 13:08:35.091604 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:08:35 crc kubenswrapper[4578]: I1003 13:08:35.092183 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.260934 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6c675fb79f-8sdmv"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.262248 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-8sdmv" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.268051 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-mhthw" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.273397 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79d68d6c85-jb6mh"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.274285 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-jb6mh" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.276747 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-xjsm6" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.286033 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6c675fb79f-8sdmv"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.296507 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79d68d6c85-jb6mh"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.304779 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-x8jtc"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.306253 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x8jtc" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.314832 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-wr9gz" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.330545 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-846dff85b5-q5cs2"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.331427 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-q5cs2" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.336114 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-z44h7" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.341496 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-x8jtc"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.341857 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnlg2\" (UniqueName: \"kubernetes.io/projected/2d6b7a4e-dd13-443e-ac6b-bc0882a0a773-kube-api-access-fnlg2\") pod \"cinder-operator-controller-manager-79d68d6c85-jb6mh\" (UID: \"2d6b7a4e-dd13-443e-ac6b-bc0882a0a773\") " pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-jb6mh" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.341919 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj9w8\" (UniqueName: \"kubernetes.io/projected/b9f0875d-37d8-4486-8618-eb0ff333f1d5-kube-api-access-kj9w8\") pod \"barbican-operator-controller-manager-6c675fb79f-8sdmv\" (UID: \"b9f0875d-37d8-4486-8618-eb0ff333f1d5\") " pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-8sdmv" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.350959 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-599898f689-vjh6c"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.352471 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-599898f689-vjh6c" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.358482 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-ndj9s" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.370819 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-846dff85b5-q5cs2"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.404807 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-599898f689-vjh6c"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.419051 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6769b867d9-dt8s5"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.427531 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-dt8s5" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.429725 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6769b867d9-dt8s5"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.464092 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.465085 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnlg2\" (UniqueName: \"kubernetes.io/projected/2d6b7a4e-dd13-443e-ac6b-bc0882a0a773-kube-api-access-fnlg2\") pod \"cinder-operator-controller-manager-79d68d6c85-jb6mh\" (UID: \"2d6b7a4e-dd13-443e-ac6b-bc0882a0a773\") " pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-jb6mh" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.465146 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmj4r\" (UniqueName: \"kubernetes.io/projected/86ba4ceb-a98b-473f-a133-c20b0d95ca04-kube-api-access-jmj4r\") pod \"glance-operator-controller-manager-846dff85b5-q5cs2\" (UID: \"86ba4ceb-a98b-473f-a133-c20b0d95ca04\") " pod="openstack-operators/glance-operator-controller-manager-846dff85b5-q5cs2" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.465192 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj9w8\" (UniqueName: \"kubernetes.io/projected/b9f0875d-37d8-4486-8618-eb0ff333f1d5-kube-api-access-kj9w8\") pod \"barbican-operator-controller-manager-6c675fb79f-8sdmv\" (UID: \"b9f0875d-37d8-4486-8618-eb0ff333f1d5\") " pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-8sdmv" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.465226 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2zfd\" (UniqueName: \"kubernetes.io/projected/311cfa36-937f-4531-ab2e-f4dec6164051-kube-api-access-x2zfd\") pod \"heat-operator-controller-manager-599898f689-vjh6c\" (UID: \"311cfa36-937f-4531-ab2e-f4dec6164051\") " pod="openstack-operators/heat-operator-controller-manager-599898f689-vjh6c" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.465273 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbs6l\" (UniqueName: \"kubernetes.io/projected/b1fc462d-d689-4a5a-b0cb-d71ea7cb8f76-kube-api-access-bbs6l\") pod \"designate-operator-controller-manager-75dfd9b554-x8jtc\" (UID: \"b1fc462d-d689-4a5a-b0cb-d71ea7cb8f76\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x8jtc" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.475341 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-svhkp" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.489159 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.493010 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-84bc9db6cc-2f2dh"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.498559 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-2f2dh" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.504496 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7f55849f88-92tt9"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.518330 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-92tt9" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.522436 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.557906 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-dmvkz" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.558682 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-m4qp2" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.558800 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.559206 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-hmlbx" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.575787 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2zfd\" (UniqueName: \"kubernetes.io/projected/311cfa36-937f-4531-ab2e-f4dec6164051-kube-api-access-x2zfd\") pod \"heat-operator-controller-manager-599898f689-vjh6c\" (UID: \"311cfa36-937f-4531-ab2e-f4dec6164051\") " pod="openstack-operators/heat-operator-controller-manager-599898f689-vjh6c" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.575841 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbs6l\" (UniqueName: \"kubernetes.io/projected/b1fc462d-d689-4a5a-b0cb-d71ea7cb8f76-kube-api-access-bbs6l\") pod \"designate-operator-controller-manager-75dfd9b554-x8jtc\" (UID: \"b1fc462d-d689-4a5a-b0cb-d71ea7cb8f76\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x8jtc" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.575866 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6644n\" (UniqueName: \"kubernetes.io/projected/a517e8b1-3fa5-41f6-a6a5-d22d9f341201-kube-api-access-6644n\") pod \"infra-operator-controller-manager-5fbf469cd7-d6mqh\" (UID: \"a517e8b1-3fa5-41f6-a6a5-d22d9f341201\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.575909 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwzw6\" (UniqueName: \"kubernetes.io/projected/86baff4f-cbb5-470f-919c-d8b571a8befb-kube-api-access-vwzw6\") pod \"horizon-operator-controller-manager-6769b867d9-dt8s5\" (UID: \"86baff4f-cbb5-470f-919c-d8b571a8befb\") " pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-dt8s5" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.575935 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a517e8b1-3fa5-41f6-a6a5-d22d9f341201-cert\") pod \"infra-operator-controller-manager-5fbf469cd7-d6mqh\" (UID: \"a517e8b1-3fa5-41f6-a6a5-d22d9f341201\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.575952 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmj4r\" (UniqueName: \"kubernetes.io/projected/86ba4ceb-a98b-473f-a133-c20b0d95ca04-kube-api-access-jmj4r\") pod \"glance-operator-controller-manager-846dff85b5-q5cs2\" (UID: \"86ba4ceb-a98b-473f-a133-c20b0d95ca04\") " pod="openstack-operators/glance-operator-controller-manager-846dff85b5-q5cs2" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.590678 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-84bc9db6cc-2f2dh"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.600983 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj9w8\" (UniqueName: \"kubernetes.io/projected/b9f0875d-37d8-4486-8618-eb0ff333f1d5-kube-api-access-kj9w8\") pod \"barbican-operator-controller-manager-6c675fb79f-8sdmv\" (UID: \"b9f0875d-37d8-4486-8618-eb0ff333f1d5\") " pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-8sdmv" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.605865 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnlg2\" (UniqueName: \"kubernetes.io/projected/2d6b7a4e-dd13-443e-ac6b-bc0882a0a773-kube-api-access-fnlg2\") pod \"cinder-operator-controller-manager-79d68d6c85-jb6mh\" (UID: \"2d6b7a4e-dd13-443e-ac6b-bc0882a0a773\") " pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-jb6mh" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.610132 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-jb6mh" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.617878 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6fd6854b49-kgt86"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.618906 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-kgt86" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.626666 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-65p76" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.648331 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2zfd\" (UniqueName: \"kubernetes.io/projected/311cfa36-937f-4531-ab2e-f4dec6164051-kube-api-access-x2zfd\") pod \"heat-operator-controller-manager-599898f689-vjh6c\" (UID: \"311cfa36-937f-4531-ab2e-f4dec6164051\") " pod="openstack-operators/heat-operator-controller-manager-599898f689-vjh6c" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.649797 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmj4r\" (UniqueName: \"kubernetes.io/projected/86ba4ceb-a98b-473f-a133-c20b0d95ca04-kube-api-access-jmj4r\") pod \"glance-operator-controller-manager-846dff85b5-q5cs2\" (UID: \"86ba4ceb-a98b-473f-a133-c20b0d95ca04\") " pod="openstack-operators/glance-operator-controller-manager-846dff85b5-q5cs2" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.657306 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7f55849f88-92tt9"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.666679 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-q5cs2" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.672218 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbs6l\" (UniqueName: \"kubernetes.io/projected/b1fc462d-d689-4a5a-b0cb-d71ea7cb8f76-kube-api-access-bbs6l\") pod \"designate-operator-controller-manager-75dfd9b554-x8jtc\" (UID: \"b1fc462d-d689-4a5a-b0cb-d71ea7cb8f76\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x8jtc" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.680737 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-cn6bd"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.681768 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-cn6bd" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.683331 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwzw6\" (UniqueName: \"kubernetes.io/projected/86baff4f-cbb5-470f-919c-d8b571a8befb-kube-api-access-vwzw6\") pod \"horizon-operator-controller-manager-6769b867d9-dt8s5\" (UID: \"86baff4f-cbb5-470f-919c-d8b571a8befb\") " pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-dt8s5" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.683483 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a517e8b1-3fa5-41f6-a6a5-d22d9f341201-cert\") pod \"infra-operator-controller-manager-5fbf469cd7-d6mqh\" (UID: \"a517e8b1-3fa5-41f6-a6a5-d22d9f341201\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.683660 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkxd8\" (UniqueName: \"kubernetes.io/projected/1e453068-d56e-4085-85fa-3bbaadd48c13-kube-api-access-dkxd8\") pod \"keystone-operator-controller-manager-7f55849f88-92tt9\" (UID: \"1e453068-d56e-4085-85fa-3bbaadd48c13\") " pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-92tt9" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.683794 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6644n\" (UniqueName: \"kubernetes.io/projected/a517e8b1-3fa5-41f6-a6a5-d22d9f341201-kube-api-access-6644n\") pod \"infra-operator-controller-manager-5fbf469cd7-d6mqh\" (UID: \"a517e8b1-3fa5-41f6-a6a5-d22d9f341201\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.683899 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vklrv\" (UniqueName: \"kubernetes.io/projected/f6d68890-a075-4dca-b4cf-19325838c18e-kube-api-access-vklrv\") pod \"ironic-operator-controller-manager-84bc9db6cc-2f2dh\" (UID: \"f6d68890-a075-4dca-b4cf-19325838c18e\") " pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-2f2dh" Oct 03 13:08:45 crc kubenswrapper[4578]: E1003 13:08:45.684365 4578 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 03 13:08:45 crc kubenswrapper[4578]: E1003 13:08:45.684532 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a517e8b1-3fa5-41f6-a6a5-d22d9f341201-cert podName:a517e8b1-3fa5-41f6-a6a5-d22d9f341201 nodeName:}" failed. No retries permitted until 2025-10-03 13:08:46.184489805 +0000 UTC m=+1061.982961989 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a517e8b1-3fa5-41f6-a6a5-d22d9f341201-cert") pod "infra-operator-controller-manager-5fbf469cd7-d6mqh" (UID: "a517e8b1-3fa5-41f6-a6a5-d22d9f341201") : secret "infra-operator-webhook-server-cert" not found Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.685321 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-h94ls" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.686110 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-599898f689-vjh6c" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.688713 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6fd6854b49-kgt86"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.712922 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-cn6bd"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.760664 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6574bf987d-s54bp"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.761945 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-s54bp" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.777845 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-555c7456bd-p4chq"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.779381 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6644n\" (UniqueName: \"kubernetes.io/projected/a517e8b1-3fa5-41f6-a6a5-d22d9f341201-kube-api-access-6644n\") pod \"infra-operator-controller-manager-5fbf469cd7-d6mqh\" (UID: \"a517e8b1-3fa5-41f6-a6a5-d22d9f341201\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.779894 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-p6gp5" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.780529 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwzw6\" (UniqueName: \"kubernetes.io/projected/86baff4f-cbb5-470f-919c-d8b571a8befb-kube-api-access-vwzw6\") pod \"horizon-operator-controller-manager-6769b867d9-dt8s5\" (UID: \"86baff4f-cbb5-470f-919c-d8b571a8befb\") " pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-dt8s5" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.786880 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-p4chq" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.794684 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szv4p\" (UniqueName: \"kubernetes.io/projected/efbb9c56-e481-4f3e-b657-a97bd2952eee-kube-api-access-szv4p\") pod \"manila-operator-controller-manager-6fd6854b49-kgt86\" (UID: \"efbb9c56-e481-4f3e-b657-a97bd2952eee\") " pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-kgt86" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.794762 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkxd8\" (UniqueName: \"kubernetes.io/projected/1e453068-d56e-4085-85fa-3bbaadd48c13-kube-api-access-dkxd8\") pod \"keystone-operator-controller-manager-7f55849f88-92tt9\" (UID: \"1e453068-d56e-4085-85fa-3bbaadd48c13\") " pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-92tt9" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.794802 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vklrv\" (UniqueName: \"kubernetes.io/projected/f6d68890-a075-4dca-b4cf-19325838c18e-kube-api-access-vklrv\") pod \"ironic-operator-controller-manager-84bc9db6cc-2f2dh\" (UID: \"f6d68890-a075-4dca-b4cf-19325838c18e\") " pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-2f2dh" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.794911 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bwvm\" (UniqueName: \"kubernetes.io/projected/e252fd2f-b318-4292-b359-7b42ca159c26-kube-api-access-9bwvm\") pod \"mariadb-operator-controller-manager-5c468bf4d4-cn6bd\" (UID: \"e252fd2f-b318-4292-b359-7b42ca159c26\") " pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-cn6bd" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.802086 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-cd66h" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.802302 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6574bf987d-s54bp"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.809726 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-59d6cfdf45-b28kv"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.810677 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-b28kv" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.823647 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-lk6c5" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.841750 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-555c7456bd-p4chq"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.860909 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-59d6cfdf45-b28kv"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.879504 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vklrv\" (UniqueName: \"kubernetes.io/projected/f6d68890-a075-4dca-b4cf-19325838c18e-kube-api-access-vklrv\") pod \"ironic-operator-controller-manager-84bc9db6cc-2f2dh\" (UID: \"f6d68890-a075-4dca-b4cf-19325838c18e\") " pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-2f2dh" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.886929 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkxd8\" (UniqueName: \"kubernetes.io/projected/1e453068-d56e-4085-85fa-3bbaadd48c13-kube-api-access-dkxd8\") pod \"keystone-operator-controller-manager-7f55849f88-92tt9\" (UID: \"1e453068-d56e-4085-85fa-3bbaadd48c13\") " pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-92tt9" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.898427 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bwvm\" (UniqueName: \"kubernetes.io/projected/e252fd2f-b318-4292-b359-7b42ca159c26-kube-api-access-9bwvm\") pod \"mariadb-operator-controller-manager-5c468bf4d4-cn6bd\" (UID: \"e252fd2f-b318-4292-b359-7b42ca159c26\") " pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-cn6bd" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.898472 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p88bk\" (UniqueName: \"kubernetes.io/projected/5848e7b5-08c6-4a7d-9fe9-3021f8aa0ac8-kube-api-access-p88bk\") pod \"nova-operator-controller-manager-555c7456bd-p4chq\" (UID: \"5848e7b5-08c6-4a7d-9fe9-3021f8aa0ac8\") " pod="openstack-operators/nova-operator-controller-manager-555c7456bd-p4chq" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.898499 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szv4p\" (UniqueName: \"kubernetes.io/projected/efbb9c56-e481-4f3e-b657-a97bd2952eee-kube-api-access-szv4p\") pod \"manila-operator-controller-manager-6fd6854b49-kgt86\" (UID: \"efbb9c56-e481-4f3e-b657-a97bd2952eee\") " pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-kgt86" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.898527 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dknb\" (UniqueName: \"kubernetes.io/projected/249acba6-7e64-4f9f-8f60-22740371de34-kube-api-access-2dknb\") pod \"neutron-operator-controller-manager-6574bf987d-s54bp\" (UID: \"249acba6-7e64-4f9f-8f60-22740371de34\") " pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-s54bp" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.900695 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-8sdmv" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.901353 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.902519 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.911477 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.921477 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-688db7b6c7-6ft9g"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.925748 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-6ft9g" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.927103 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-4cfr8" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.937848 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-lhmmj" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.945149 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x8jtc" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.952731 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-688db7b6c7-6ft9g"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.964966 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.972054 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-7d8bb7f44c-ql5zc"] Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.978243 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-ql5zc" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.984201 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szv4p\" (UniqueName: \"kubernetes.io/projected/efbb9c56-e481-4f3e-b657-a97bd2952eee-kube-api-access-szv4p\") pod \"manila-operator-controller-manager-6fd6854b49-kgt86\" (UID: \"efbb9c56-e481-4f3e-b657-a97bd2952eee\") " pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-kgt86" Oct 03 13:08:45 crc kubenswrapper[4578]: I1003 13:08:45.991989 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-7d8bb7f44c-ql5zc"] Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.002028 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-kzr6z"] Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.003111 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kzr6z" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.004072 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p88bk\" (UniqueName: \"kubernetes.io/projected/5848e7b5-08c6-4a7d-9fe9-3021f8aa0ac8-kube-api-access-p88bk\") pod \"nova-operator-controller-manager-555c7456bd-p4chq\" (UID: \"5848e7b5-08c6-4a7d-9fe9-3021f8aa0ac8\") " pod="openstack-operators/nova-operator-controller-manager-555c7456bd-p4chq" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.004128 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9f9f\" (UniqueName: \"kubernetes.io/projected/a152b6b9-86b3-489d-9ceb-a9a284e054d1-kube-api-access-k9f9f\") pod \"octavia-operator-controller-manager-59d6cfdf45-b28kv\" (UID: \"a152b6b9-86b3-489d-9ceb-a9a284e054d1\") " pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-b28kv" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.004161 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dknb\" (UniqueName: \"kubernetes.io/projected/249acba6-7e64-4f9f-8f60-22740371de34-kube-api-access-2dknb\") pod \"neutron-operator-controller-manager-6574bf987d-s54bp\" (UID: \"249acba6-7e64-4f9f-8f60-22740371de34\") " pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-s54bp" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.004192 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx96t\" (UniqueName: \"kubernetes.io/projected/31250d8b-75f5-4145-b8ed-1ec79bebe936-kube-api-access-wx96t\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d6789s9md\" (UID: \"31250d8b-75f5-4145-b8ed-1ec79bebe936\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.004283 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31250d8b-75f5-4145-b8ed-1ec79bebe936-cert\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d6789s9md\" (UID: \"31250d8b-75f5-4145-b8ed-1ec79bebe936\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.006364 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bwvm\" (UniqueName: \"kubernetes.io/projected/e252fd2f-b318-4292-b359-7b42ca159c26-kube-api-access-9bwvm\") pod \"mariadb-operator-controller-manager-5c468bf4d4-cn6bd\" (UID: \"e252fd2f-b318-4292-b359-7b42ca159c26\") " pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-cn6bd" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.006492 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-xrnvj" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.012860 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5db5cf686f-rpzgw"] Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.014696 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-bsjqw" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.016792 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-rpzgw" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.029430 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-kzr6z"] Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.037172 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-bccb6" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.047746 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5db5cf686f-rpzgw"] Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.052021 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-2f2dh" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.057787 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-r2zgb"] Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.070783 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p88bk\" (UniqueName: \"kubernetes.io/projected/5848e7b5-08c6-4a7d-9fe9-3021f8aa0ac8-kube-api-access-p88bk\") pod \"nova-operator-controller-manager-555c7456bd-p4chq\" (UID: \"5848e7b5-08c6-4a7d-9fe9-3021f8aa0ac8\") " pod="openstack-operators/nova-operator-controller-manager-555c7456bd-p4chq" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.076659 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-dt8s5" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.079124 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-r2zgb" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.104190 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-92tt9" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.105276 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-b28q8" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.107482 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-r2zgb"] Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.113391 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dknb\" (UniqueName: \"kubernetes.io/projected/249acba6-7e64-4f9f-8f60-22740371de34-kube-api-access-2dknb\") pod \"neutron-operator-controller-manager-6574bf987d-s54bp\" (UID: \"249acba6-7e64-4f9f-8f60-22740371de34\") " pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-s54bp" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.139083 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-kgt86" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.162652 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31250d8b-75f5-4145-b8ed-1ec79bebe936-cert\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d6789s9md\" (UID: \"31250d8b-75f5-4145-b8ed-1ec79bebe936\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.166184 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-cn6bd" Oct 03 13:08:46 crc kubenswrapper[4578]: E1003 13:08:46.175119 4578 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 03 13:08:46 crc kubenswrapper[4578]: E1003 13:08:46.175203 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31250d8b-75f5-4145-b8ed-1ec79bebe936-cert podName:31250d8b-75f5-4145-b8ed-1ec79bebe936 nodeName:}" failed. No retries permitted until 2025-10-03 13:08:46.675186013 +0000 UTC m=+1062.473658197 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/31250d8b-75f5-4145-b8ed-1ec79bebe936-cert") pod "openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" (UID: "31250d8b-75f5-4145-b8ed-1ec79bebe936") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.180384 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvqqp\" (UniqueName: \"kubernetes.io/projected/5b7d3469-bfd4-47c8-a82c-3e00721b2102-kube-api-access-qvqqp\") pod \"test-operator-controller-manager-5cd5cb47d7-r2zgb\" (UID: \"5b7d3469-bfd4-47c8-a82c-3e00721b2102\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-r2zgb" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.180506 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnvmc\" (UniqueName: \"kubernetes.io/projected/65e1c76d-7bb5-4c09-8b8b-d54f3d92c42b-kube-api-access-bnvmc\") pod \"swift-operator-controller-manager-6859f9b676-kzr6z\" (UID: \"65e1c76d-7bb5-4c09-8b8b-d54f3d92c42b\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kzr6z" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.180564 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9f9f\" (UniqueName: \"kubernetes.io/projected/a152b6b9-86b3-489d-9ceb-a9a284e054d1-kube-api-access-k9f9f\") pod \"octavia-operator-controller-manager-59d6cfdf45-b28kv\" (UID: \"a152b6b9-86b3-489d-9ceb-a9a284e054d1\") " pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-b28kv" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.180720 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx96t\" (UniqueName: \"kubernetes.io/projected/31250d8b-75f5-4145-b8ed-1ec79bebe936-kube-api-access-wx96t\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d6789s9md\" (UID: \"31250d8b-75f5-4145-b8ed-1ec79bebe936\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.180792 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29r57\" (UniqueName: \"kubernetes.io/projected/48ea8320-2f7a-46e5-8681-73d075a6b74b-kube-api-access-29r57\") pod \"ovn-operator-controller-manager-688db7b6c7-6ft9g\" (UID: \"48ea8320-2f7a-46e5-8681-73d075a6b74b\") " pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-6ft9g" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.180890 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8tcn\" (UniqueName: \"kubernetes.io/projected/d8bb1592-732e-4e22-944e-3298b6937f55-kube-api-access-n8tcn\") pod \"placement-operator-controller-manager-7d8bb7f44c-ql5zc\" (UID: \"d8bb1592-732e-4e22-944e-3298b6937f55\") " pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-ql5zc" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.186879 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-s54bp" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.217525 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx96t\" (UniqueName: \"kubernetes.io/projected/31250d8b-75f5-4145-b8ed-1ec79bebe936-kube-api-access-wx96t\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d6789s9md\" (UID: \"31250d8b-75f5-4145-b8ed-1ec79bebe936\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.223970 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-fcd7d9895-dcr4b"] Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.224436 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-p4chq" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.226776 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-dcr4b" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.247054 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9f9f\" (UniqueName: \"kubernetes.io/projected/a152b6b9-86b3-489d-9ceb-a9a284e054d1-kube-api-access-k9f9f\") pod \"octavia-operator-controller-manager-59d6cfdf45-b28kv\" (UID: \"a152b6b9-86b3-489d-9ceb-a9a284e054d1\") " pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-b28kv" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.249101 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-fhfsl" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.280214 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-b28kv" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.283314 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5msjg\" (UniqueName: \"kubernetes.io/projected/4a1657ed-b0b2-4f52-9f9c-38ed6d6bb8ab-kube-api-access-5msjg\") pod \"watcher-operator-controller-manager-fcd7d9895-dcr4b\" (UID: \"4a1657ed-b0b2-4f52-9f9c-38ed6d6bb8ab\") " pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-dcr4b" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.283358 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29r57\" (UniqueName: \"kubernetes.io/projected/48ea8320-2f7a-46e5-8681-73d075a6b74b-kube-api-access-29r57\") pod \"ovn-operator-controller-manager-688db7b6c7-6ft9g\" (UID: \"48ea8320-2f7a-46e5-8681-73d075a6b74b\") " pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-6ft9g" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.283395 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8tcn\" (UniqueName: \"kubernetes.io/projected/d8bb1592-732e-4e22-944e-3298b6937f55-kube-api-access-n8tcn\") pod \"placement-operator-controller-manager-7d8bb7f44c-ql5zc\" (UID: \"d8bb1592-732e-4e22-944e-3298b6937f55\") " pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-ql5zc" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.283425 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a517e8b1-3fa5-41f6-a6a5-d22d9f341201-cert\") pod \"infra-operator-controller-manager-5fbf469cd7-d6mqh\" (UID: \"a517e8b1-3fa5-41f6-a6a5-d22d9f341201\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.283448 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb68t\" (UniqueName: \"kubernetes.io/projected/1a74246b-d89b-424f-bd47-d57674b06855-kube-api-access-lb68t\") pod \"telemetry-operator-controller-manager-5db5cf686f-rpzgw\" (UID: \"1a74246b-d89b-424f-bd47-d57674b06855\") " pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-rpzgw" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.283491 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvqqp\" (UniqueName: \"kubernetes.io/projected/5b7d3469-bfd4-47c8-a82c-3e00721b2102-kube-api-access-qvqqp\") pod \"test-operator-controller-manager-5cd5cb47d7-r2zgb\" (UID: \"5b7d3469-bfd4-47c8-a82c-3e00721b2102\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-r2zgb" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.283514 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnvmc\" (UniqueName: \"kubernetes.io/projected/65e1c76d-7bb5-4c09-8b8b-d54f3d92c42b-kube-api-access-bnvmc\") pod \"swift-operator-controller-manager-6859f9b676-kzr6z\" (UID: \"65e1c76d-7bb5-4c09-8b8b-d54f3d92c42b\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kzr6z" Oct 03 13:08:46 crc kubenswrapper[4578]: E1003 13:08:46.283846 4578 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 03 13:08:46 crc kubenswrapper[4578]: E1003 13:08:46.283894 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a517e8b1-3fa5-41f6-a6a5-d22d9f341201-cert podName:a517e8b1-3fa5-41f6-a6a5-d22d9f341201 nodeName:}" failed. No retries permitted until 2025-10-03 13:08:47.283878507 +0000 UTC m=+1063.082350691 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a517e8b1-3fa5-41f6-a6a5-d22d9f341201-cert") pod "infra-operator-controller-manager-5fbf469cd7-d6mqh" (UID: "a517e8b1-3fa5-41f6-a6a5-d22d9f341201") : secret "infra-operator-webhook-server-cert" not found Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.297382 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-fcd7d9895-dcr4b"] Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.352369 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvqqp\" (UniqueName: \"kubernetes.io/projected/5b7d3469-bfd4-47c8-a82c-3e00721b2102-kube-api-access-qvqqp\") pod \"test-operator-controller-manager-5cd5cb47d7-r2zgb\" (UID: \"5b7d3469-bfd4-47c8-a82c-3e00721b2102\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-r2zgb" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.352997 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnvmc\" (UniqueName: \"kubernetes.io/projected/65e1c76d-7bb5-4c09-8b8b-d54f3d92c42b-kube-api-access-bnvmc\") pod \"swift-operator-controller-manager-6859f9b676-kzr6z\" (UID: \"65e1c76d-7bb5-4c09-8b8b-d54f3d92c42b\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kzr6z" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.354535 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-r2zgb" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.371164 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8tcn\" (UniqueName: \"kubernetes.io/projected/d8bb1592-732e-4e22-944e-3298b6937f55-kube-api-access-n8tcn\") pod \"placement-operator-controller-manager-7d8bb7f44c-ql5zc\" (UID: \"d8bb1592-732e-4e22-944e-3298b6937f55\") " pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-ql5zc" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.383229 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29r57\" (UniqueName: \"kubernetes.io/projected/48ea8320-2f7a-46e5-8681-73d075a6b74b-kube-api-access-29r57\") pod \"ovn-operator-controller-manager-688db7b6c7-6ft9g\" (UID: \"48ea8320-2f7a-46e5-8681-73d075a6b74b\") " pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-6ft9g" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.408846 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-ql5zc" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.410982 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb68t\" (UniqueName: \"kubernetes.io/projected/1a74246b-d89b-424f-bd47-d57674b06855-kube-api-access-lb68t\") pod \"telemetry-operator-controller-manager-5db5cf686f-rpzgw\" (UID: \"1a74246b-d89b-424f-bd47-d57674b06855\") " pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-rpzgw" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.411082 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5msjg\" (UniqueName: \"kubernetes.io/projected/4a1657ed-b0b2-4f52-9f9c-38ed6d6bb8ab-kube-api-access-5msjg\") pod \"watcher-operator-controller-manager-fcd7d9895-dcr4b\" (UID: \"4a1657ed-b0b2-4f52-9f9c-38ed6d6bb8ab\") " pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-dcr4b" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.444730 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5c4446bf96-nxg6g"] Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.445956 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5c4446bf96-nxg6g" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.455301 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-7p5b4" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.455821 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.476142 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5c4446bf96-nxg6g"] Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.504620 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb68t\" (UniqueName: \"kubernetes.io/projected/1a74246b-d89b-424f-bd47-d57674b06855-kube-api-access-lb68t\") pod \"telemetry-operator-controller-manager-5db5cf686f-rpzgw\" (UID: \"1a74246b-d89b-424f-bd47-d57674b06855\") " pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-rpzgw" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.515325 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx5xg\" (UniqueName: \"kubernetes.io/projected/4855be16-97a1-467a-bfbe-36b2f15a5afd-kube-api-access-nx5xg\") pod \"openstack-operator-controller-manager-5c4446bf96-nxg6g\" (UID: \"4855be16-97a1-467a-bfbe-36b2f15a5afd\") " pod="openstack-operators/openstack-operator-controller-manager-5c4446bf96-nxg6g" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.515373 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4855be16-97a1-467a-bfbe-36b2f15a5afd-cert\") pod \"openstack-operator-controller-manager-5c4446bf96-nxg6g\" (UID: \"4855be16-97a1-467a-bfbe-36b2f15a5afd\") " pod="openstack-operators/openstack-operator-controller-manager-5c4446bf96-nxg6g" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.534594 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5msjg\" (UniqueName: \"kubernetes.io/projected/4a1657ed-b0b2-4f52-9f9c-38ed6d6bb8ab-kube-api-access-5msjg\") pod \"watcher-operator-controller-manager-fcd7d9895-dcr4b\" (UID: \"4a1657ed-b0b2-4f52-9f9c-38ed6d6bb8ab\") " pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-dcr4b" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.560251 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kzr6z" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.600983 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-rpzgw" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.633984 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx5xg\" (UniqueName: \"kubernetes.io/projected/4855be16-97a1-467a-bfbe-36b2f15a5afd-kube-api-access-nx5xg\") pod \"openstack-operator-controller-manager-5c4446bf96-nxg6g\" (UID: \"4855be16-97a1-467a-bfbe-36b2f15a5afd\") " pod="openstack-operators/openstack-operator-controller-manager-5c4446bf96-nxg6g" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.634034 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4855be16-97a1-467a-bfbe-36b2f15a5afd-cert\") pod \"openstack-operator-controller-manager-5c4446bf96-nxg6g\" (UID: \"4855be16-97a1-467a-bfbe-36b2f15a5afd\") " pod="openstack-operators/openstack-operator-controller-manager-5c4446bf96-nxg6g" Oct 03 13:08:46 crc kubenswrapper[4578]: E1003 13:08:46.634213 4578 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 03 13:08:46 crc kubenswrapper[4578]: E1003 13:08:46.634259 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4855be16-97a1-467a-bfbe-36b2f15a5afd-cert podName:4855be16-97a1-467a-bfbe-36b2f15a5afd nodeName:}" failed. No retries permitted until 2025-10-03 13:08:47.134245214 +0000 UTC m=+1062.932717388 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4855be16-97a1-467a-bfbe-36b2f15a5afd-cert") pod "openstack-operator-controller-manager-5c4446bf96-nxg6g" (UID: "4855be16-97a1-467a-bfbe-36b2f15a5afd") : secret "webhook-server-cert" not found Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.664930 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-mj2n5"] Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.665785 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-mj2n5" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.669699 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-6ft9g" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.698156 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx5xg\" (UniqueName: \"kubernetes.io/projected/4855be16-97a1-467a-bfbe-36b2f15a5afd-kube-api-access-nx5xg\") pod \"openstack-operator-controller-manager-5c4446bf96-nxg6g\" (UID: \"4855be16-97a1-467a-bfbe-36b2f15a5afd\") " pod="openstack-operators/openstack-operator-controller-manager-5c4446bf96-nxg6g" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.701787 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-mj2n5"] Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.705239 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-x9mc4" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.699626 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-dcr4b" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.739554 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31250d8b-75f5-4145-b8ed-1ec79bebe936-cert\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d6789s9md\" (UID: \"31250d8b-75f5-4145-b8ed-1ec79bebe936\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.739657 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7fcp\" (UniqueName: \"kubernetes.io/projected/1141b204-4370-4c24-bbf3-71734e237805-kube-api-access-q7fcp\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-mj2n5\" (UID: \"1141b204-4370-4c24-bbf3-71734e237805\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-mj2n5" Oct 03 13:08:46 crc kubenswrapper[4578]: E1003 13:08:46.739816 4578 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 03 13:08:46 crc kubenswrapper[4578]: E1003 13:08:46.739857 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31250d8b-75f5-4145-b8ed-1ec79bebe936-cert podName:31250d8b-75f5-4145-b8ed-1ec79bebe936 nodeName:}" failed. No retries permitted until 2025-10-03 13:08:47.739841361 +0000 UTC m=+1063.538313545 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/31250d8b-75f5-4145-b8ed-1ec79bebe936-cert") pod "openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" (UID: "31250d8b-75f5-4145-b8ed-1ec79bebe936") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.741517 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79d68d6c85-jb6mh"] Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.841566 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7fcp\" (UniqueName: \"kubernetes.io/projected/1141b204-4370-4c24-bbf3-71734e237805-kube-api-access-q7fcp\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-mj2n5\" (UID: \"1141b204-4370-4c24-bbf3-71734e237805\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-mj2n5" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.875578 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7fcp\" (UniqueName: \"kubernetes.io/projected/1141b204-4370-4c24-bbf3-71734e237805-kube-api-access-q7fcp\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-mj2n5\" (UID: \"1141b204-4370-4c24-bbf3-71734e237805\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-mj2n5" Oct 03 13:08:46 crc kubenswrapper[4578]: I1003 13:08:46.894350 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-mj2n5" Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.005039 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-599898f689-vjh6c"] Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.145820 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4855be16-97a1-467a-bfbe-36b2f15a5afd-cert\") pod \"openstack-operator-controller-manager-5c4446bf96-nxg6g\" (UID: \"4855be16-97a1-467a-bfbe-36b2f15a5afd\") " pod="openstack-operators/openstack-operator-controller-manager-5c4446bf96-nxg6g" Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.153615 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4855be16-97a1-467a-bfbe-36b2f15a5afd-cert\") pod \"openstack-operator-controller-manager-5c4446bf96-nxg6g\" (UID: \"4855be16-97a1-467a-bfbe-36b2f15a5afd\") " pod="openstack-operators/openstack-operator-controller-manager-5c4446bf96-nxg6g" Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.304964 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-846dff85b5-q5cs2"] Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.317484 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-jb6mh" event={"ID":"2d6b7a4e-dd13-443e-ac6b-bc0882a0a773","Type":"ContainerStarted","Data":"05e6ecc7a99f1565a31e8de1d93a6b4a29b9eaf337aeb0dfc37fd355a524d3fe"} Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.317774 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6c675fb79f-8sdmv"] Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.319682 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-599898f689-vjh6c" event={"ID":"311cfa36-937f-4531-ab2e-f4dec6164051","Type":"ContainerStarted","Data":"c06472279ba358b2e204ee3ce188c423b53ea876456aa3b50848a27d862a3e84"} Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.350740 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a517e8b1-3fa5-41f6-a6a5-d22d9f341201-cert\") pod \"infra-operator-controller-manager-5fbf469cd7-d6mqh\" (UID: \"a517e8b1-3fa5-41f6-a6a5-d22d9f341201\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh" Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.369152 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a517e8b1-3fa5-41f6-a6a5-d22d9f341201-cert\") pod \"infra-operator-controller-manager-5fbf469cd7-d6mqh\" (UID: \"a517e8b1-3fa5-41f6-a6a5-d22d9f341201\") " pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh" Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.400277 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5c4446bf96-nxg6g" Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.425182 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh" Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.631749 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6574bf987d-s54bp"] Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.652931 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7f55849f88-92tt9"] Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.671281 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-x8jtc"] Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.716701 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-84bc9db6cc-2f2dh"] Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.738625 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6769b867d9-dt8s5"] Oct 03 13:08:47 crc kubenswrapper[4578]: W1003 13:08:47.752200 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86baff4f_cbb5_470f_919c_d8b571a8befb.slice/crio-450a2fb0bdde224baa3793c00c23585e741afaffe74a132449ff4f4b5a0c42ca WatchSource:0}: Error finding container 450a2fb0bdde224baa3793c00c23585e741afaffe74a132449ff4f4b5a0c42ca: Status 404 returned error can't find the container with id 450a2fb0bdde224baa3793c00c23585e741afaffe74a132449ff4f4b5a0c42ca Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.756281 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31250d8b-75f5-4145-b8ed-1ec79bebe936-cert\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d6789s9md\" (UID: \"31250d8b-75f5-4145-b8ed-1ec79bebe936\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.766186 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31250d8b-75f5-4145-b8ed-1ec79bebe936-cert\") pod \"openstack-baremetal-operator-controller-manager-6f64c4d6789s9md\" (UID: \"31250d8b-75f5-4145-b8ed-1ec79bebe936\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" Oct 03 13:08:47 crc kubenswrapper[4578]: W1003 13:08:47.784375 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6d68890_a075_4dca_b4cf_19325838c18e.slice/crio-7c19aa703f25f77d358796d22c840804f02f733d6acb7c314c0a2e037c560c44 WatchSource:0}: Error finding container 7c19aa703f25f77d358796d22c840804f02f733d6acb7c314c0a2e037c560c44: Status 404 returned error can't find the container with id 7c19aa703f25f77d358796d22c840804f02f733d6acb7c314c0a2e037c560c44 Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.831577 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.929138 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-fcd7d9895-dcr4b"] Oct 03 13:08:47 crc kubenswrapper[4578]: W1003 13:08:47.965978 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a1657ed_b0b2_4f52_9f9c_38ed6d6bb8ab.slice/crio-63463f18ad5000e0532ae09768ee5e4ff6042549062bec226520fb1ba393f517 WatchSource:0}: Error finding container 63463f18ad5000e0532ae09768ee5e4ff6042549062bec226520fb1ba393f517: Status 404 returned error can't find the container with id 63463f18ad5000e0532ae09768ee5e4ff6042549062bec226520fb1ba393f517 Oct 03 13:08:47 crc kubenswrapper[4578]: I1003 13:08:47.988569 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-cn6bd"] Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.118707 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-mj2n5"] Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.125672 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-7d8bb7f44c-ql5zc"] Oct 03 13:08:48 crc kubenswrapper[4578]: W1003 13:08:48.126341 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1141b204_4370_4c24_bbf3_71734e237805.slice/crio-efbd566486aa94c9f661103b723300e853102e7058b7e068258cfb355cc11808 WatchSource:0}: Error finding container efbd566486aa94c9f661103b723300e853102e7058b7e068258cfb355cc11808: Status 404 returned error can't find the container with id efbd566486aa94c9f661103b723300e853102e7058b7e068258cfb355cc11808 Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.128135 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-kzr6z"] Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.138751 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5db5cf686f-rpzgw"] Oct 03 13:08:48 crc kubenswrapper[4578]: W1003 13:08:48.143357 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65e1c76d_7bb5_4c09_8b8b_d54f3d92c42b.slice/crio-4e6f6c8b30d94d0b936192aa154119a7b8af9d51ce19d3c5e3e9db3d0f037c81 WatchSource:0}: Error finding container 4e6f6c8b30d94d0b936192aa154119a7b8af9d51ce19d3c5e3e9db3d0f037c81: Status 404 returned error can't find the container with id 4e6f6c8b30d94d0b936192aa154119a7b8af9d51ce19d3c5e3e9db3d0f037c81 Oct 03 13:08:48 crc kubenswrapper[4578]: W1003 13:08:48.143529 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8bb1592_732e_4e22_944e_3298b6937f55.slice/crio-35a0f1823c7070d772adf54139619fcd2538c8de9882981cf6733ddb826722c9 WatchSource:0}: Error finding container 35a0f1823c7070d772adf54139619fcd2538c8de9882981cf6733ddb826722c9: Status 404 returned error can't find the container with id 35a0f1823c7070d772adf54139619fcd2538c8de9882981cf6733ddb826722c9 Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.158130 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6fd6854b49-kgt86"] Oct 03 13:08:48 crc kubenswrapper[4578]: W1003 13:08:48.174104 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5848e7b5_08c6_4a7d_9fe9_3021f8aa0ac8.slice/crio-0d97ee925e9229593660bffdd955f4eaccae1b503ee445b8745e434e41116053 WatchSource:0}: Error finding container 0d97ee925e9229593660bffdd955f4eaccae1b503ee445b8745e434e41116053: Status 404 returned error can't find the container with id 0d97ee925e9229593660bffdd955f4eaccae1b503ee445b8745e434e41116053 Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.183524 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-555c7456bd-p4chq"] Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.190590 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-688db7b6c7-6ft9g"] Oct 03 13:08:48 crc kubenswrapper[4578]: E1003 13:08:48.193988 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bnvmc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6859f9b676-kzr6z_openstack-operators(65e1c76d-7bb5-4c09-8b8b-d54f3d92c42b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 03 13:08:48 crc kubenswrapper[4578]: E1003 13:08:48.194336 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:5c6ab93b78bd20eb7f1736751a59c1eb33fb06351339563dbefe49ccaaff6e94,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-29r57,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-688db7b6c7-6ft9g_openstack-operators(48ea8320-2f7a-46e5-8681-73d075a6b74b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.294487 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-59d6cfdf45-b28kv"] Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.317450 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh"] Oct 03 13:08:48 crc kubenswrapper[4578]: W1003 13:08:48.331310 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda517e8b1_3fa5_41f6_a6a5_d22d9f341201.slice/crio-7da280521c541ce0bc05697c94fe02ae4bea3032dc78c591a1c5ff9aaa66e3e5 WatchSource:0}: Error finding container 7da280521c541ce0bc05697c94fe02ae4bea3032dc78c591a1c5ff9aaa66e3e5: Status 404 returned error can't find the container with id 7da280521c541ce0bc05697c94fe02ae4bea3032dc78c591a1c5ff9aaa66e3e5 Oct 03 13:08:48 crc kubenswrapper[4578]: W1003 13:08:48.334807 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda152b6b9_86b3_489d_9ceb_a9a284e054d1.slice/crio-3cd1aa57db066a6d78efb663aea42d8851d791033194ee9fda8251bc73592165 WatchSource:0}: Error finding container 3cd1aa57db066a6d78efb663aea42d8851d791033194ee9fda8251bc73592165: Status 404 returned error can't find the container with id 3cd1aa57db066a6d78efb663aea42d8851d791033194ee9fda8251bc73592165 Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.334910 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-kgt86" event={"ID":"efbb9c56-e481-4f3e-b657-a97bd2952eee","Type":"ContainerStarted","Data":"f9c86b941a41528bcb2d69c3fc11406d4646b52de865023372c429386dae2ab0"} Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.339156 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-dt8s5" event={"ID":"86baff4f-cbb5-470f-919c-d8b571a8befb","Type":"ContainerStarted","Data":"450a2fb0bdde224baa3793c00c23585e741afaffe74a132449ff4f4b5a0c42ca"} Oct 03 13:08:48 crc kubenswrapper[4578]: E1003 13:08:48.340197 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:40fb1819b6639807b77ef79448d35f1e4bfc1838a09d4f380e9fa0f755352475,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6644n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-5fbf469cd7-d6mqh_openstack-operators(a517e8b1-3fa5-41f6-a6a5-d22d9f341201): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 03 13:08:48 crc kubenswrapper[4578]: E1003 13:08:48.341617 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:b7409dcf05c85eab205904d29d4276f8e927c772eba6363ecfa21ab10c4aaa01,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k9f9f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-59d6cfdf45-b28kv_openstack-operators(a152b6b9-86b3-489d-9ceb-a9a284e054d1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.349004 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-2f2dh" event={"ID":"f6d68890-a075-4dca-b4cf-19325838c18e","Type":"ContainerStarted","Data":"7c19aa703f25f77d358796d22c840804f02f733d6acb7c314c0a2e037c560c44"} Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.373017 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-6ft9g" event={"ID":"48ea8320-2f7a-46e5-8681-73d075a6b74b","Type":"ContainerStarted","Data":"df652db3e2ca67fdeed37895a78bc981038de24a7356c5963c0c35782ce0a496"} Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.376094 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-cn6bd" event={"ID":"e252fd2f-b318-4292-b359-7b42ca159c26","Type":"ContainerStarted","Data":"d368bbb9e7a56e696fd2051949a15aebc43338ead077baf03aac20375503efc8"} Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.379972 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-q5cs2" event={"ID":"86ba4ceb-a98b-473f-a133-c20b0d95ca04","Type":"ContainerStarted","Data":"8dd298a946054f08822d09e1f65bb9a946f65eb71e15f29d3a3b6dab75ccdbde"} Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.382549 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kzr6z" event={"ID":"65e1c76d-7bb5-4c09-8b8b-d54f3d92c42b","Type":"ContainerStarted","Data":"4e6f6c8b30d94d0b936192aa154119a7b8af9d51ce19d3c5e3e9db3d0f037c81"} Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.384660 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-92tt9" event={"ID":"1e453068-d56e-4085-85fa-3bbaadd48c13","Type":"ContainerStarted","Data":"a4eef76008431235e38b1c1178da8f884a5af0a657cc3606f75a2063db866edc"} Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.386080 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-rpzgw" event={"ID":"1a74246b-d89b-424f-bd47-d57674b06855","Type":"ContainerStarted","Data":"6ea73563963a870b4bca554ca87f8860c95dc7e6e72a79144200e3a641e2cdfe"} Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.387376 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-dcr4b" event={"ID":"4a1657ed-b0b2-4f52-9f9c-38ed6d6bb8ab","Type":"ContainerStarted","Data":"63463f18ad5000e0532ae09768ee5e4ff6042549062bec226520fb1ba393f517"} Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.389351 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-mj2n5" event={"ID":"1141b204-4370-4c24-bbf3-71734e237805","Type":"ContainerStarted","Data":"efbd566486aa94c9f661103b723300e853102e7058b7e068258cfb355cc11808"} Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.390917 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x8jtc" event={"ID":"b1fc462d-d689-4a5a-b0cb-d71ea7cb8f76","Type":"ContainerStarted","Data":"1477b982b6f89276b25e30fb3aa28c804f50b12881b4a2b6fd928b97ccc1b281"} Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.392238 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-p4chq" event={"ID":"5848e7b5-08c6-4a7d-9fe9-3021f8aa0ac8","Type":"ContainerStarted","Data":"0d97ee925e9229593660bffdd955f4eaccae1b503ee445b8745e434e41116053"} Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.393268 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-8sdmv" event={"ID":"b9f0875d-37d8-4486-8618-eb0ff333f1d5","Type":"ContainerStarted","Data":"9501cd340fcb31eaafb2f71c07be5b720099feba41eeac673a9284cbfd1a39e8"} Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.394295 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-ql5zc" event={"ID":"d8bb1592-732e-4e22-944e-3298b6937f55","Type":"ContainerStarted","Data":"35a0f1823c7070d772adf54139619fcd2538c8de9882981cf6733ddb826722c9"} Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.395500 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-s54bp" event={"ID":"249acba6-7e64-4f9f-8f60-22740371de34","Type":"ContainerStarted","Data":"57efced8da7c899020a7e6ad6c11a1ec51abd9d6f0eb819d16070b857fec1940"} Oct 03 13:08:48 crc kubenswrapper[4578]: E1003 13:08:48.493714 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-6ft9g" podUID="48ea8320-2f7a-46e5-8681-73d075a6b74b" Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.499495 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-r2zgb"] Oct 03 13:08:48 crc kubenswrapper[4578]: E1003 13:08:48.505775 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kzr6z" podUID="65e1c76d-7bb5-4c09-8b8b-d54f3d92c42b" Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.510131 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5c4446bf96-nxg6g"] Oct 03 13:08:48 crc kubenswrapper[4578]: W1003 13:08:48.537056 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4855be16_97a1_467a_bfbe_36b2f15a5afd.slice/crio-07eab4fc7966b51a4f54e9a2a1c142ac520457f96457ba0a00c4dba0198c79d3 WatchSource:0}: Error finding container 07eab4fc7966b51a4f54e9a2a1c142ac520457f96457ba0a00c4dba0198c79d3: Status 404 returned error can't find the container with id 07eab4fc7966b51a4f54e9a2a1c142ac520457f96457ba0a00c4dba0198c79d3 Oct 03 13:08:48 crc kubenswrapper[4578]: I1003 13:08:48.552857 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md"] Oct 03 13:08:48 crc kubenswrapper[4578]: W1003 13:08:48.572680 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b7d3469_bfd4_47c8_a82c_3e00721b2102.slice/crio-7769364aadbe889eb529905cb39f132590b9ebbe69d3fae33e2aa3d363091294 WatchSource:0}: Error finding container 7769364aadbe889eb529905cb39f132590b9ebbe69d3fae33e2aa3d363091294: Status 404 returned error can't find the container with id 7769364aadbe889eb529905cb39f132590b9ebbe69d3fae33e2aa3d363091294 Oct 03 13:08:48 crc kubenswrapper[4578]: W1003 13:08:48.588048 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31250d8b_75f5_4145_b8ed_1ec79bebe936.slice/crio-14925c1635ed0d8de9bb226f873b2664a393e7ffa8f6a3ce82d54f86712fda2b WatchSource:0}: Error finding container 14925c1635ed0d8de9bb226f873b2664a393e7ffa8f6a3ce82d54f86712fda2b: Status 404 returned error can't find the container with id 14925c1635ed0d8de9bb226f873b2664a393e7ffa8f6a3ce82d54f86712fda2b Oct 03 13:08:48 crc kubenswrapper[4578]: E1003 13:08:48.713113 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-b28kv" podUID="a152b6b9-86b3-489d-9ceb-a9a284e054d1" Oct 03 13:08:48 crc kubenswrapper[4578]: E1003 13:08:48.719138 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh" podUID="a517e8b1-3fa5-41f6-a6a5-d22d9f341201" Oct 03 13:08:49 crc kubenswrapper[4578]: I1003 13:08:49.484764 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-6ft9g" event={"ID":"48ea8320-2f7a-46e5-8681-73d075a6b74b","Type":"ContainerStarted","Data":"69006268735a95bfea23cb1791dfa317a9ceb8127b1d25d215a304ae985e40b4"} Oct 03 13:08:49 crc kubenswrapper[4578]: E1003 13:08:49.490287 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5c6ab93b78bd20eb7f1736751a59c1eb33fb06351339563dbefe49ccaaff6e94\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-6ft9g" podUID="48ea8320-2f7a-46e5-8681-73d075a6b74b" Oct 03 13:08:49 crc kubenswrapper[4578]: I1003 13:08:49.492013 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-r2zgb" event={"ID":"5b7d3469-bfd4-47c8-a82c-3e00721b2102","Type":"ContainerStarted","Data":"7769364aadbe889eb529905cb39f132590b9ebbe69d3fae33e2aa3d363091294"} Oct 03 13:08:49 crc kubenswrapper[4578]: I1003 13:08:49.495019 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5c4446bf96-nxg6g" event={"ID":"4855be16-97a1-467a-bfbe-36b2f15a5afd","Type":"ContainerStarted","Data":"51d238d21641107ca3c735ae42623dd047ccdecaba7b64c1d0cfa58a13909846"} Oct 03 13:08:49 crc kubenswrapper[4578]: I1003 13:08:49.495046 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5c4446bf96-nxg6g" event={"ID":"4855be16-97a1-467a-bfbe-36b2f15a5afd","Type":"ContainerStarted","Data":"8a2c48c2ba1d401a10b3a5f534cc669031805a10313cae4bbe10ad806f787c28"} Oct 03 13:08:49 crc kubenswrapper[4578]: I1003 13:08:49.495055 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5c4446bf96-nxg6g" event={"ID":"4855be16-97a1-467a-bfbe-36b2f15a5afd","Type":"ContainerStarted","Data":"07eab4fc7966b51a4f54e9a2a1c142ac520457f96457ba0a00c4dba0198c79d3"} Oct 03 13:08:49 crc kubenswrapper[4578]: I1003 13:08:49.495268 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5c4446bf96-nxg6g" Oct 03 13:08:49 crc kubenswrapper[4578]: I1003 13:08:49.501022 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh" event={"ID":"a517e8b1-3fa5-41f6-a6a5-d22d9f341201","Type":"ContainerStarted","Data":"aa0802d0e083de7ccc6c7461a8bdc1b55bdd8b59af8ae5923fefd2e914cd1044"} Oct 03 13:08:49 crc kubenswrapper[4578]: I1003 13:08:49.501083 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh" event={"ID":"a517e8b1-3fa5-41f6-a6a5-d22d9f341201","Type":"ContainerStarted","Data":"7da280521c541ce0bc05697c94fe02ae4bea3032dc78c591a1c5ff9aaa66e3e5"} Oct 03 13:08:49 crc kubenswrapper[4578]: E1003 13:08:49.507329 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:40fb1819b6639807b77ef79448d35f1e4bfc1838a09d4f380e9fa0f755352475\\\"\"" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh" podUID="a517e8b1-3fa5-41f6-a6a5-d22d9f341201" Oct 03 13:08:49 crc kubenswrapper[4578]: I1003 13:08:49.513439 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-b28kv" event={"ID":"a152b6b9-86b3-489d-9ceb-a9a284e054d1","Type":"ContainerStarted","Data":"be842d1de7d49cf4c3a94878def6a2d4762f13f8a3ce9f1984d543c04c9eb5ce"} Oct 03 13:08:49 crc kubenswrapper[4578]: I1003 13:08:49.513481 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-b28kv" event={"ID":"a152b6b9-86b3-489d-9ceb-a9a284e054d1","Type":"ContainerStarted","Data":"3cd1aa57db066a6d78efb663aea42d8851d791033194ee9fda8251bc73592165"} Oct 03 13:08:49 crc kubenswrapper[4578]: E1003 13:08:49.521659 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:b7409dcf05c85eab205904d29d4276f8e927c772eba6363ecfa21ab10c4aaa01\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-b28kv" podUID="a152b6b9-86b3-489d-9ceb-a9a284e054d1" Oct 03 13:08:49 crc kubenswrapper[4578]: I1003 13:08:49.549258 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" event={"ID":"31250d8b-75f5-4145-b8ed-1ec79bebe936","Type":"ContainerStarted","Data":"14925c1635ed0d8de9bb226f873b2664a393e7ffa8f6a3ce82d54f86712fda2b"} Oct 03 13:08:49 crc kubenswrapper[4578]: I1003 13:08:49.562651 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kzr6z" event={"ID":"65e1c76d-7bb5-4c09-8b8b-d54f3d92c42b","Type":"ContainerStarted","Data":"7b7b4c003ffbe72b784bddfcdb3d838c7ca5c64ab9f4a39b594625d1d2e8799d"} Oct 03 13:08:49 crc kubenswrapper[4578]: I1003 13:08:49.564849 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5c4446bf96-nxg6g" podStartSLOduration=3.564822805 podStartE2EDuration="3.564822805s" podCreationTimestamp="2025-10-03 13:08:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:08:49.551298079 +0000 UTC m=+1065.349770263" watchObservedRunningTime="2025-10-03 13:08:49.564822805 +0000 UTC m=+1065.363294989" Oct 03 13:08:49 crc kubenswrapper[4578]: E1003 13:08:49.571029 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed\\\"\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kzr6z" podUID="65e1c76d-7bb5-4c09-8b8b-d54f3d92c42b" Oct 03 13:08:50 crc kubenswrapper[4578]: E1003 13:08:50.590903 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:40fb1819b6639807b77ef79448d35f1e4bfc1838a09d4f380e9fa0f755352475\\\"\"" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh" podUID="a517e8b1-3fa5-41f6-a6a5-d22d9f341201" Oct 03 13:08:50 crc kubenswrapper[4578]: E1003 13:08:50.592029 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:b7409dcf05c85eab205904d29d4276f8e927c772eba6363ecfa21ab10c4aaa01\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-b28kv" podUID="a152b6b9-86b3-489d-9ceb-a9a284e054d1" Oct 03 13:08:50 crc kubenswrapper[4578]: E1003 13:08:50.592060 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5c6ab93b78bd20eb7f1736751a59c1eb33fb06351339563dbefe49ccaaff6e94\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-6ft9g" podUID="48ea8320-2f7a-46e5-8681-73d075a6b74b" Oct 03 13:08:50 crc kubenswrapper[4578]: E1003 13:08:50.593015 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed\\\"\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kzr6z" podUID="65e1c76d-7bb5-4c09-8b8b-d54f3d92c42b" Oct 03 13:08:57 crc kubenswrapper[4578]: I1003 13:08:57.406866 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5c4446bf96-nxg6g" Oct 03 13:09:02 crc kubenswrapper[4578]: E1003 13:09:02.522353 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:38abe6135ccaa369bc831f7878a6dfdf9a5a993a882e1c42073ca43582766f12" Oct 03 13:09:02 crc kubenswrapper[4578]: E1003 13:09:02.523094 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:38abe6135ccaa369bc831f7878a6dfdf9a5a993a882e1c42073ca43582766f12,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vklrv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-84bc9db6cc-2f2dh_openstack-operators(f6d68890-a075-4dca-b4cf-19325838c18e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:09:03 crc kubenswrapper[4578]: E1003 13:09:03.235173 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:a82409e6d6a5554aad95acfe6fa4784e33de19a963eb8b1da1a80a3e6cf1ab55" Oct 03 13:09:03 crc kubenswrapper[4578]: E1003 13:09:03.235405 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:a82409e6d6a5554aad95acfe6fa4784e33de19a963eb8b1da1a80a3e6cf1ab55,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-p88bk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-555c7456bd-p4chq_openstack-operators(5848e7b5-08c6-4a7d-9fe9-3021f8aa0ac8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:09:04 crc kubenswrapper[4578]: E1003 13:09:04.205459 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:e9ff0784bffe5b9a6d1a77a1b8866dd26b8d0c54465707df1808f68caad93a95" Oct 03 13:09:04 crc kubenswrapper[4578]: E1003 13:09:04.205683 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:e9ff0784bffe5b9a6d1a77a1b8866dd26b8d0c54465707df1808f68caad93a95,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vwzw6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-6769b867d9-dt8s5_openstack-operators(86baff4f-cbb5-470f-919c-d8b571a8befb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:09:04 crc kubenswrapper[4578]: E1003 13:09:04.667235 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:110b885fe640ffdd8536e7da2a613677a6777e3d902e2ff15fa4d5968fe06c54" Oct 03 13:09:04 crc kubenswrapper[4578]: E1003 13:09:04.667800 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:110b885fe640ffdd8536e7da2a613677a6777e3d902e2ff15fa4d5968fe06c54,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9bwvm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-5c468bf4d4-cn6bd_openstack-operators(e252fd2f-b318-4292-b359-7b42ca159c26): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:09:05 crc kubenswrapper[4578]: E1003 13:09:05.090050 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:1bd15383c53b5772acd12c376bd29b9fde0c2e6cca2cd493492fb1eeb9c52ef5" Oct 03 13:09:05 crc kubenswrapper[4578]: E1003 13:09:05.090993 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:1bd15383c53b5772acd12c376bd29b9fde0c2e6cca2cd493492fb1eeb9c52ef5,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x2zfd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-599898f689-vjh6c_openstack-operators(311cfa36-937f-4531-ab2e-f4dec6164051): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:09:05 crc kubenswrapper[4578]: I1003 13:09:05.091522 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:09:05 crc kubenswrapper[4578]: I1003 13:09:05.091578 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:09:05 crc kubenswrapper[4578]: E1003 13:09:05.505077 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:8f5eee2eb7b77432ef1a88ed693ff981514359dfc808581f393bcef252de5cfa" Oct 03 13:09:05 crc kubenswrapper[4578]: E1003 13:09:05.505249 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:8f5eee2eb7b77432ef1a88ed693ff981514359dfc808581f393bcef252de5cfa,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lb68t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5db5cf686f-rpzgw_openstack-operators(1a74246b-d89b-424f-bd47-d57674b06855): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:09:08 crc kubenswrapper[4578]: E1003 13:09:08.638507 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:8fdb7ea8542adb2eca73f11bd78e6aebceed2ba7a1e9fdd149c75e0049d09ce0" Oct 03 13:09:08 crc kubenswrapper[4578]: E1003 13:09:08.639300 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:8fdb7ea8542adb2eca73f11bd78e6aebceed2ba7a1e9fdd149c75e0049d09ce0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-szv4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-6fd6854b49-kgt86_openstack-operators(efbb9c56-e481-4f3e-b657-a97bd2952eee): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:09:09 crc kubenswrapper[4578]: E1003 13:09:09.067127 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:018151bd5ff830ec03c6b8e3d53cfb9456ca6e1e34793bdd4f7edd39a0146fa6" Oct 03 13:09:09 crc kubenswrapper[4578]: E1003 13:09:09.067344 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:018151bd5ff830ec03c6b8e3d53cfb9456ca6e1e34793bdd4f7edd39a0146fa6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5msjg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-fcd7d9895-dcr4b_openstack-operators(4a1657ed-b0b2-4f52-9f9c-38ed6d6bb8ab): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:09:10 crc kubenswrapper[4578]: E1003 13:09:10.582269 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:f50229c8a33fd581bccbe5f34bbaf3936c1b454802e755c9b48b40b76a8239ee" Oct 03 13:09:10 crc kubenswrapper[4578]: E1003 13:09:10.584818 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:f50229c8a33fd581bccbe5f34bbaf3936c1b454802e755c9b48b40b76a8239ee,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wx96t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-6f64c4d6789s9md_openstack-operators(31250d8b-75f5-4145-b8ed-1ec79bebe936): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:09:11 crc kubenswrapper[4578]: E1003 13:09:11.205377 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:570e59f91d7dd66c9abcec1e54889a44c65d676d3fff6802be101fe5215bc988" Oct 03 13:09:11 crc kubenswrapper[4578]: E1003 13:09:11.205626 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:570e59f91d7dd66c9abcec1e54889a44c65d676d3fff6802be101fe5215bc988,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2dknb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-6574bf987d-s54bp_openstack-operators(249acba6-7e64-4f9f-8f60-22740371de34): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:09:11 crc kubenswrapper[4578]: E1003 13:09:11.644878 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb" Oct 03 13:09:11 crc kubenswrapper[4578]: E1003 13:09:11.645367 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qvqqp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd5cb47d7-r2zgb_openstack-operators(5b7d3469-bfd4-47c8-a82c-3e00721b2102): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:09:12 crc kubenswrapper[4578]: E1003 13:09:12.052348 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:9fed055cd1f09627ef351e61c7e42227570193ccd5d33167a607c49b442a9d87" Oct 03 13:09:12 crc kubenswrapper[4578]: E1003 13:09:12.052573 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:9fed055cd1f09627ef351e61c7e42227570193ccd5d33167a607c49b442a9d87,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jmj4r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-846dff85b5-q5cs2_openstack-operators(86ba4ceb-a98b-473f-a133-c20b0d95ca04): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:09:18 crc kubenswrapper[4578]: E1003 13:09:18.138043 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-p4chq" podUID="5848e7b5-08c6-4a7d-9fe9-3021f8aa0ac8" Oct 03 13:09:18 crc kubenswrapper[4578]: E1003 13:09:18.291909 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-dt8s5" podUID="86baff4f-cbb5-470f-919c-d8b571a8befb" Oct 03 13:09:18 crc kubenswrapper[4578]: E1003 13:09:18.562165 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-dcr4b" podUID="4a1657ed-b0b2-4f52-9f9c-38ed6d6bb8ab" Oct 03 13:09:18 crc kubenswrapper[4578]: E1003 13:09:18.565194 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-kgt86" podUID="efbb9c56-e481-4f3e-b657-a97bd2952eee" Oct 03 13:09:18 crc kubenswrapper[4578]: E1003 13:09:18.565333 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-2f2dh" podUID="f6d68890-a075-4dca-b4cf-19325838c18e" Oct 03 13:09:18 crc kubenswrapper[4578]: E1003 13:09:18.565433 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-rpzgw" podUID="1a74246b-d89b-424f-bd47-d57674b06855" Oct 03 13:09:18 crc kubenswrapper[4578]: E1003 13:09:18.565574 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-cn6bd" podUID="e252fd2f-b318-4292-b359-7b42ca159c26" Oct 03 13:09:18 crc kubenswrapper[4578]: E1003 13:09:18.628417 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" podUID="31250d8b-75f5-4145-b8ed-1ec79bebe936" Oct 03 13:09:18 crc kubenswrapper[4578]: E1003 13:09:18.675050 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-599898f689-vjh6c" podUID="311cfa36-937f-4531-ab2e-f4dec6164051" Oct 03 13:09:18 crc kubenswrapper[4578]: E1003 13:09:18.682416 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-q5cs2" podUID="86ba4ceb-a98b-473f-a133-c20b0d95ca04" Oct 03 13:09:18 crc kubenswrapper[4578]: E1003 13:09:18.782938 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-r2zgb" podUID="5b7d3469-bfd4-47c8-a82c-3e00721b2102" Oct 03 13:09:18 crc kubenswrapper[4578]: I1003 13:09:18.822391 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-kgt86" event={"ID":"efbb9c56-e481-4f3e-b657-a97bd2952eee","Type":"ContainerStarted","Data":"735ab8211e9abbccff1a66914f7d76ac24e1e8c17c0246ab30424667d5d00058"} Oct 03 13:09:18 crc kubenswrapper[4578]: I1003 13:09:18.829308 4578 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 13:09:18 crc kubenswrapper[4578]: I1003 13:09:18.830836 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-dt8s5" event={"ID":"86baff4f-cbb5-470f-919c-d8b571a8befb","Type":"ContainerStarted","Data":"c659a2fe1b4934f03cb6696b4b8716d6f0cea7adf3c780322476fca95589b0b8"} Oct 03 13:09:18 crc kubenswrapper[4578]: I1003 13:09:18.840788 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-r2zgb" event={"ID":"5b7d3469-bfd4-47c8-a82c-3e00721b2102","Type":"ContainerStarted","Data":"8a36bd6fc51570926474e055b2a74825f3b43f47594a00f556fa9c8b59999c25"} Oct 03 13:09:18 crc kubenswrapper[4578]: E1003 13:09:18.843860 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-r2zgb" podUID="5b7d3469-bfd4-47c8-a82c-3e00721b2102" Oct 03 13:09:18 crc kubenswrapper[4578]: I1003 13:09:18.851922 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-dcr4b" event={"ID":"4a1657ed-b0b2-4f52-9f9c-38ed6d6bb8ab","Type":"ContainerStarted","Data":"b99ec36ae30bf3675684fac4a640ecaab3ac3cccd21e49bc32fbe46e478a5174"} Oct 03 13:09:18 crc kubenswrapper[4578]: E1003 13:09:18.855519 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:018151bd5ff830ec03c6b8e3d53cfb9456ca6e1e34793bdd4f7edd39a0146fa6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-dcr4b" podUID="4a1657ed-b0b2-4f52-9f9c-38ed6d6bb8ab" Oct 03 13:09:18 crc kubenswrapper[4578]: I1003 13:09:18.862048 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-cn6bd" event={"ID":"e252fd2f-b318-4292-b359-7b42ca159c26","Type":"ContainerStarted","Data":"7b7d6ad1597e3141ad5a2f2bd882aecb7778c471af0b22af19f265cc81c8c9f8"} Oct 03 13:09:18 crc kubenswrapper[4578]: I1003 13:09:18.874703 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-rpzgw" event={"ID":"1a74246b-d89b-424f-bd47-d57674b06855","Type":"ContainerStarted","Data":"5993260f1dceb1c23b3a2014b4ad4899d7742ef7c9558e79b2c8f954697d12a8"} Oct 03 13:09:18 crc kubenswrapper[4578]: I1003 13:09:18.889430 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-2f2dh" event={"ID":"f6d68890-a075-4dca-b4cf-19325838c18e","Type":"ContainerStarted","Data":"289e14427c71f80e179adb2463bb395d55f5f473b69a5770685cdcf4382306fa"} Oct 03 13:09:18 crc kubenswrapper[4578]: I1003 13:09:18.971504 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-599898f689-vjh6c" event={"ID":"311cfa36-937f-4531-ab2e-f4dec6164051","Type":"ContainerStarted","Data":"7df7282f095fd2445c0a83d63a7e9ae9e42a88802717f2fb880fe00af64753a3"} Oct 03 13:09:18 crc kubenswrapper[4578]: I1003 13:09:18.972469 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" event={"ID":"31250d8b-75f5-4145-b8ed-1ec79bebe936","Type":"ContainerStarted","Data":"e5c2d0551e9e9c4060de8217d0815c8075dca1ab8f9774d5f8a6634893f145a3"} Oct 03 13:09:18 crc kubenswrapper[4578]: E1003 13:09:18.981345 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:f50229c8a33fd581bccbe5f34bbaf3936c1b454802e755c9b48b40b76a8239ee\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" podUID="31250d8b-75f5-4145-b8ed-1ec79bebe936" Oct 03 13:09:18 crc kubenswrapper[4578]: E1003 13:09:18.986329 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-s54bp" podUID="249acba6-7e64-4f9f-8f60-22740371de34" Oct 03 13:09:18 crc kubenswrapper[4578]: I1003 13:09:18.999209 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-p4chq" event={"ID":"5848e7b5-08c6-4a7d-9fe9-3021f8aa0ac8","Type":"ContainerStarted","Data":"a509f0a6157cd42ad59e2a24a3d359e789793d6ee536effa7e7b063f436c6ffc"} Oct 03 13:09:19 crc kubenswrapper[4578]: I1003 13:09:19.014545 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-q5cs2" event={"ID":"86ba4ceb-a98b-473f-a133-c20b0d95ca04","Type":"ContainerStarted","Data":"3557c8919201210fb403614f383352950832a06693aa1f18ba23ae98efbc609d"} Oct 03 13:09:19 crc kubenswrapper[4578]: E1003 13:09:19.019015 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:9fed055cd1f09627ef351e61c7e42227570193ccd5d33167a607c49b442a9d87\\\"\"" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-q5cs2" podUID="86ba4ceb-a98b-473f-a133-c20b0d95ca04" Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.033956 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh" event={"ID":"a517e8b1-3fa5-41f6-a6a5-d22d9f341201","Type":"ContainerStarted","Data":"2b3b1050eee16f0eb429014b306336ff6284cf140dbdeccd7668e553c11dd776"} Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.036404 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh" Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.038082 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-b28kv" event={"ID":"a152b6b9-86b3-489d-9ceb-a9a284e054d1","Type":"ContainerStarted","Data":"a4f16e054339f9c82b7f75b7e6fc5a6b68cb46ebbf14f34881515292feaea5bc"} Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.038809 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-b28kv" Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.042145 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-jb6mh" event={"ID":"2d6b7a4e-dd13-443e-ac6b-bc0882a0a773","Type":"ContainerStarted","Data":"f961d516fe68fc79503ed79c2b7e184c91c5c0f1b7531fe31d3e24c9b068dda6"} Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.042330 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-jb6mh" event={"ID":"2d6b7a4e-dd13-443e-ac6b-bc0882a0a773","Type":"ContainerStarted","Data":"9d77ff32a5b0e4cdbaae4bb8ef61da5d9bc8714f9c3ab5a301fee923ae43da37"} Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.043154 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-jb6mh" Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.047048 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-ql5zc" event={"ID":"d8bb1592-732e-4e22-944e-3298b6937f55","Type":"ContainerStarted","Data":"7b9cae8eb7b3d513a0452fc583560a9e08c8cb01c69c95b057b0615dc7b789dd"} Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.058176 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kzr6z" event={"ID":"65e1c76d-7bb5-4c09-8b8b-d54f3d92c42b","Type":"ContainerStarted","Data":"05af6a759ea2170321ac2e7b7aa14d10b533c97a1f812783447b7d51c86d03c7"} Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.058904 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kzr6z" Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.063915 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-s54bp" event={"ID":"249acba6-7e64-4f9f-8f60-22740371de34","Type":"ContainerStarted","Data":"58fae78dc6e080ce930e04f37a4cd4d4e6950c0296f9bd344f999f1781e32b3c"} Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.069154 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-mj2n5" event={"ID":"1141b204-4370-4c24-bbf3-71734e237805","Type":"ContainerStarted","Data":"1356cd176d20cf0212debebf28193216cb53469c9fe205dada234dd016777616"} Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.072624 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x8jtc" event={"ID":"b1fc462d-d689-4a5a-b0cb-d71ea7cb8f76","Type":"ContainerStarted","Data":"7c3b6ce65fc6927ad2ed3f1aa669dbd9c186f9365c198a16f814da0f3df956d7"} Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.072911 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x8jtc" event={"ID":"b1fc462d-d689-4a5a-b0cb-d71ea7cb8f76","Type":"ContainerStarted","Data":"9470cc925dcd3781db92ca2ace8f6d5f22b9b538f76b4d9562731882bb68c758"} Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.073133 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x8jtc" Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.080492 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-92tt9" event={"ID":"1e453068-d56e-4085-85fa-3bbaadd48c13","Type":"ContainerStarted","Data":"f2930a821886bbeb37078d7bdde26912178339201f7f2ba85e15857057813d8b"} Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.091572 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-8sdmv" event={"ID":"b9f0875d-37d8-4486-8618-eb0ff333f1d5","Type":"ContainerStarted","Data":"beb1ba261f2179c01bdd58376f565c96e0787d1f81bf7955df1ed076f9a5edff"} Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.091615 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-8sdmv" event={"ID":"b9f0875d-37d8-4486-8618-eb0ff333f1d5","Type":"ContainerStarted","Data":"c88fa6aac30c2ffc443fef9f418077c720a61a7a40cb5a0b28d50999df0ebe81"} Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.091865 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-8sdmv" Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.104261 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-6ft9g" event={"ID":"48ea8320-2f7a-46e5-8681-73d075a6b74b","Type":"ContainerStarted","Data":"317702be155b9750a7e0c544d3df9a7270465c6f57291466b7dddaed0c778d36"} Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.129395 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh" podStartSLOduration=5.194490943 podStartE2EDuration="35.129381145s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:48.339426412 +0000 UTC m=+1064.137898596" lastFinishedPulling="2025-10-03 13:09:18.274316624 +0000 UTC m=+1094.072788798" observedRunningTime="2025-10-03 13:09:20.125060978 +0000 UTC m=+1095.923533162" watchObservedRunningTime="2025-10-03 13:09:20.129381145 +0000 UTC m=+1095.927853329" Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.177175 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-6ft9g" podStartSLOduration=5.12966433 podStartE2EDuration="35.1771542s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:48.194261589 +0000 UTC m=+1063.992733773" lastFinishedPulling="2025-10-03 13:09:18.241751459 +0000 UTC m=+1094.040223643" observedRunningTime="2025-10-03 13:09:20.174011501 +0000 UTC m=+1095.972483695" watchObservedRunningTime="2025-10-03 13:09:20.1771542 +0000 UTC m=+1095.975626384" Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.242769 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-mj2n5" podStartSLOduration=4.355321318 podStartE2EDuration="34.242752896s" podCreationTimestamp="2025-10-03 13:08:46 +0000 UTC" firstStartedPulling="2025-10-03 13:08:48.141654952 +0000 UTC m=+1063.940127136" lastFinishedPulling="2025-10-03 13:09:18.02908652 +0000 UTC m=+1093.827558714" observedRunningTime="2025-10-03 13:09:20.241864988 +0000 UTC m=+1096.040337172" watchObservedRunningTime="2025-10-03 13:09:20.242752896 +0000 UTC m=+1096.041225080" Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.293373 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kzr6z" podStartSLOduration=5.151118976 podStartE2EDuration="35.29335272s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:48.193870897 +0000 UTC m=+1063.992343081" lastFinishedPulling="2025-10-03 13:09:18.336104641 +0000 UTC m=+1094.134576825" observedRunningTime="2025-10-03 13:09:20.287908158 +0000 UTC m=+1096.086380342" watchObservedRunningTime="2025-10-03 13:09:20.29335272 +0000 UTC m=+1096.091824904" Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.338448 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-b28kv" podStartSLOduration=5.330047622 podStartE2EDuration="35.33843082s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:48.341545569 +0000 UTC m=+1064.140017753" lastFinishedPulling="2025-10-03 13:09:18.349928767 +0000 UTC m=+1094.148400951" observedRunningTime="2025-10-03 13:09:20.335285521 +0000 UTC m=+1096.133757715" watchObservedRunningTime="2025-10-03 13:09:20.33843082 +0000 UTC m=+1096.136903004" Oct 03 13:09:20 crc kubenswrapper[4578]: E1003 13:09:20.358576 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-r2zgb" podUID="5b7d3469-bfd4-47c8-a82c-3e00721b2102" Oct 03 13:09:20 crc kubenswrapper[4578]: E1003 13:09:20.358683 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:f50229c8a33fd581bccbe5f34bbaf3936c1b454802e755c9b48b40b76a8239ee\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" podUID="31250d8b-75f5-4145-b8ed-1ec79bebe936" Oct 03 13:09:20 crc kubenswrapper[4578]: E1003 13:09:20.358824 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:9fed055cd1f09627ef351e61c7e42227570193ccd5d33167a607c49b442a9d87\\\"\"" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-q5cs2" podUID="86ba4ceb-a98b-473f-a133-c20b0d95ca04" Oct 03 13:09:20 crc kubenswrapper[4578]: E1003 13:09:20.365721 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:570e59f91d7dd66c9abcec1e54889a44c65d676d3fff6802be101fe5215bc988\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-s54bp" podUID="249acba6-7e64-4f9f-8f60-22740371de34" Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.406847 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x8jtc" podStartSLOduration=5.091624492 podStartE2EDuration="35.406828525s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:47.708740894 +0000 UTC m=+1063.507213078" lastFinishedPulling="2025-10-03 13:09:18.023944917 +0000 UTC m=+1093.822417111" observedRunningTime="2025-10-03 13:09:20.404998447 +0000 UTC m=+1096.203470631" watchObservedRunningTime="2025-10-03 13:09:20.406828525 +0000 UTC m=+1096.205300709" Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.433861 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-8sdmv" podStartSLOduration=4.791354032 podStartE2EDuration="35.433845726s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:47.387866125 +0000 UTC m=+1063.186338309" lastFinishedPulling="2025-10-03 13:09:18.030357829 +0000 UTC m=+1093.828830003" observedRunningTime="2025-10-03 13:09:20.428849888 +0000 UTC m=+1096.227322072" watchObservedRunningTime="2025-10-03 13:09:20.433845726 +0000 UTC m=+1096.232317910" Oct 03 13:09:20 crc kubenswrapper[4578]: I1003 13:09:20.495386 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-jb6mh" podStartSLOduration=9.475875218 podStartE2EDuration="35.495372444s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:46.964145867 +0000 UTC m=+1062.762618051" lastFinishedPulling="2025-10-03 13:09:12.983643093 +0000 UTC m=+1088.782115277" observedRunningTime="2025-10-03 13:09:20.494407034 +0000 UTC m=+1096.292879218" watchObservedRunningTime="2025-10-03 13:09:20.495372444 +0000 UTC m=+1096.293844628" Oct 03 13:09:21 crc kubenswrapper[4578]: I1003 13:09:21.112461 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-kgt86" event={"ID":"efbb9c56-e481-4f3e-b657-a97bd2952eee","Type":"ContainerStarted","Data":"9dbb2854da2bd362518b69ac8011c41ea662dbe8856a7fcb7d95414b26d9c253"} Oct 03 13:09:21 crc kubenswrapper[4578]: I1003 13:09:21.112862 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-kgt86" Oct 03 13:09:21 crc kubenswrapper[4578]: I1003 13:09:21.114948 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-92tt9" event={"ID":"1e453068-d56e-4085-85fa-3bbaadd48c13","Type":"ContainerStarted","Data":"59a06086fb853c7556035d6876b0c849159b2019557e1a5b5f19f1772a6e7cdf"} Oct 03 13:09:21 crc kubenswrapper[4578]: I1003 13:09:21.115006 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-92tt9" Oct 03 13:09:21 crc kubenswrapper[4578]: I1003 13:09:21.117087 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-cn6bd" event={"ID":"e252fd2f-b318-4292-b359-7b42ca159c26","Type":"ContainerStarted","Data":"3a0a9095022011363efa22bd1425f1def76656964995798a846a0e91a26095f9"} Oct 03 13:09:21 crc kubenswrapper[4578]: I1003 13:09:21.118957 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-ql5zc" event={"ID":"d8bb1592-732e-4e22-944e-3298b6937f55","Type":"ContainerStarted","Data":"08f2f9dafcd07e43aa0e19e442760525e2fef2b8b69ab81b55ffef2ff3f5c282"} Oct 03 13:09:21 crc kubenswrapper[4578]: I1003 13:09:21.119138 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-ql5zc" Oct 03 13:09:21 crc kubenswrapper[4578]: I1003 13:09:21.122967 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-dt8s5" event={"ID":"86baff4f-cbb5-470f-919c-d8b571a8befb","Type":"ContainerStarted","Data":"2f1e2683b4234ddb3090a3bf379c598a8d9adcb148529cbd7c888ca3c1973840"} Oct 03 13:09:21 crc kubenswrapper[4578]: I1003 13:09:21.134057 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-kgt86" podStartSLOduration=4.276211134 podStartE2EDuration="36.134042284s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:48.193240757 +0000 UTC m=+1063.991712941" lastFinishedPulling="2025-10-03 13:09:20.051071907 +0000 UTC m=+1095.849544091" observedRunningTime="2025-10-03 13:09:21.13232127 +0000 UTC m=+1096.930793454" watchObservedRunningTime="2025-10-03 13:09:21.134042284 +0000 UTC m=+1096.932514478" Oct 03 13:09:21 crc kubenswrapper[4578]: E1003 13:09:21.137855 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:570e59f91d7dd66c9abcec1e54889a44c65d676d3fff6802be101fe5215bc988\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-s54bp" podUID="249acba6-7e64-4f9f-8f60-22740371de34" Oct 03 13:09:21 crc kubenswrapper[4578]: I1003 13:09:21.192848 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-ql5zc" podStartSLOduration=6.32889295 podStartE2EDuration="36.192830907s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:48.166472804 +0000 UTC m=+1063.964944988" lastFinishedPulling="2025-10-03 13:09:18.030410741 +0000 UTC m=+1093.828882945" observedRunningTime="2025-10-03 13:09:21.169974306 +0000 UTC m=+1096.968446480" watchObservedRunningTime="2025-10-03 13:09:21.192830907 +0000 UTC m=+1096.991303091" Oct 03 13:09:21 crc kubenswrapper[4578]: I1003 13:09:21.195613 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-dt8s5" podStartSLOduration=3.232473573 podStartE2EDuration="36.195603673s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:47.756659724 +0000 UTC m=+1063.555131908" lastFinishedPulling="2025-10-03 13:09:20.719789824 +0000 UTC m=+1096.518262008" observedRunningTime="2025-10-03 13:09:21.193058124 +0000 UTC m=+1096.991530308" watchObservedRunningTime="2025-10-03 13:09:21.195603673 +0000 UTC m=+1096.994075857" Oct 03 13:09:21 crc kubenswrapper[4578]: I1003 13:09:21.216463 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-92tt9" podStartSLOduration=5.874756642 podStartE2EDuration="36.216442019s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:47.688429884 +0000 UTC m=+1063.486902068" lastFinishedPulling="2025-10-03 13:09:18.030115241 +0000 UTC m=+1093.828587445" observedRunningTime="2025-10-03 13:09:21.21454697 +0000 UTC m=+1097.013019154" watchObservedRunningTime="2025-10-03 13:09:21.216442019 +0000 UTC m=+1097.014914203" Oct 03 13:09:22 crc kubenswrapper[4578]: I1003 13:09:22.130972 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-rpzgw" event={"ID":"1a74246b-d89b-424f-bd47-d57674b06855","Type":"ContainerStarted","Data":"48a7ea00132c835bac6f3121f68a07269c6671fbe3527118285a6d2ff23df807"} Oct 03 13:09:22 crc kubenswrapper[4578]: I1003 13:09:22.132141 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-rpzgw" Oct 03 13:09:22 crc kubenswrapper[4578]: I1003 13:09:22.133558 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-2f2dh" event={"ID":"f6d68890-a075-4dca-b4cf-19325838c18e","Type":"ContainerStarted","Data":"a323b7c397a4385977be6713f846c58483286b6b3cc8719bc70c72ed56c58d4a"} Oct 03 13:09:22 crc kubenswrapper[4578]: I1003 13:09:22.133934 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-2f2dh" Oct 03 13:09:22 crc kubenswrapper[4578]: I1003 13:09:22.135808 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-599898f689-vjh6c" event={"ID":"311cfa36-937f-4531-ab2e-f4dec6164051","Type":"ContainerStarted","Data":"cfa4fdd5a575b31395e6ce3864d337f6df16eeb6b3521d265948522543e68204"} Oct 03 13:09:22 crc kubenswrapper[4578]: I1003 13:09:22.136258 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-599898f689-vjh6c" Oct 03 13:09:22 crc kubenswrapper[4578]: I1003 13:09:22.139052 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-dcr4b" event={"ID":"4a1657ed-b0b2-4f52-9f9c-38ed6d6bb8ab","Type":"ContainerStarted","Data":"e85257597e2a11a1d8f37d2b0d80f21fdc8898ed37fcdae607e8c460bdca3ad3"} Oct 03 13:09:22 crc kubenswrapper[4578]: I1003 13:09:22.139449 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-dcr4b" Oct 03 13:09:22 crc kubenswrapper[4578]: I1003 13:09:22.141508 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-p4chq" event={"ID":"5848e7b5-08c6-4a7d-9fe9-3021f8aa0ac8","Type":"ContainerStarted","Data":"7ede8558019771bb5e9ca093eca942bd0b8397e3e5d12743a9d4e624deeec91d"} Oct 03 13:09:22 crc kubenswrapper[4578]: I1003 13:09:22.141563 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-cn6bd" Oct 03 13:09:22 crc kubenswrapper[4578]: I1003 13:09:22.141714 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-dt8s5" Oct 03 13:09:22 crc kubenswrapper[4578]: I1003 13:09:22.167518 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-p4chq" podStartSLOduration=4.522752301 podStartE2EDuration="37.167499331s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:48.193048661 +0000 UTC m=+1063.991520845" lastFinishedPulling="2025-10-03 13:09:20.837795691 +0000 UTC m=+1096.636267875" observedRunningTime="2025-10-03 13:09:22.167077357 +0000 UTC m=+1097.965549551" watchObservedRunningTime="2025-10-03 13:09:22.167499331 +0000 UTC m=+1097.965971515" Oct 03 13:09:22 crc kubenswrapper[4578]: I1003 13:09:22.169853 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-rpzgw" podStartSLOduration=4.628363928 podStartE2EDuration="37.169834105s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:48.19334395 +0000 UTC m=+1063.991816134" lastFinishedPulling="2025-10-03 13:09:20.734814127 +0000 UTC m=+1096.533286311" observedRunningTime="2025-10-03 13:09:22.152124367 +0000 UTC m=+1097.950596561" watchObservedRunningTime="2025-10-03 13:09:22.169834105 +0000 UTC m=+1097.968306289" Oct 03 13:09:22 crc kubenswrapper[4578]: I1003 13:09:22.208613 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-2f2dh" podStartSLOduration=4.946262333 podStartE2EDuration="37.208592856s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:47.789945492 +0000 UTC m=+1063.588417676" lastFinishedPulling="2025-10-03 13:09:20.052276015 +0000 UTC m=+1095.850748199" observedRunningTime="2025-10-03 13:09:22.205930841 +0000 UTC m=+1098.004403025" watchObservedRunningTime="2025-10-03 13:09:22.208592856 +0000 UTC m=+1098.007065040" Oct 03 13:09:22 crc kubenswrapper[4578]: I1003 13:09:22.210091 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-599898f689-vjh6c" podStartSLOduration=3.514840548 podStartE2EDuration="37.210083092s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:47.028899207 +0000 UTC m=+1062.827371391" lastFinishedPulling="2025-10-03 13:09:20.724141631 +0000 UTC m=+1096.522613935" observedRunningTime="2025-10-03 13:09:22.183231796 +0000 UTC m=+1097.981703980" watchObservedRunningTime="2025-10-03 13:09:22.210083092 +0000 UTC m=+1098.008555286" Oct 03 13:09:22 crc kubenswrapper[4578]: I1003 13:09:22.228135 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-dcr4b" podStartSLOduration=4.03383361 podStartE2EDuration="37.228114001s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:47.97587604 +0000 UTC m=+1063.774348224" lastFinishedPulling="2025-10-03 13:09:21.170156431 +0000 UTC m=+1096.968628615" observedRunningTime="2025-10-03 13:09:22.222905526 +0000 UTC m=+1098.021377730" watchObservedRunningTime="2025-10-03 13:09:22.228114001 +0000 UTC m=+1098.026586185" Oct 03 13:09:23 crc kubenswrapper[4578]: I1003 13:09:23.146409 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-p4chq" Oct 03 13:09:25 crc kubenswrapper[4578]: I1003 13:09:25.614744 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-79d68d6c85-jb6mh" Oct 03 13:09:25 crc kubenswrapper[4578]: I1003 13:09:25.633501 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-cn6bd" podStartSLOduration=7.929627496 podStartE2EDuration="40.633480939s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:48.014056742 +0000 UTC m=+1063.812528926" lastFinishedPulling="2025-10-03 13:09:20.717910185 +0000 UTC m=+1096.516382369" observedRunningTime="2025-10-03 13:09:22.250839017 +0000 UTC m=+1098.049311221" watchObservedRunningTime="2025-10-03 13:09:25.633480939 +0000 UTC m=+1101.431953143" Oct 03 13:09:25 crc kubenswrapper[4578]: I1003 13:09:25.904866 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6c675fb79f-8sdmv" Oct 03 13:09:25 crc kubenswrapper[4578]: I1003 13:09:25.949444 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-x8jtc" Oct 03 13:09:26 crc kubenswrapper[4578]: I1003 13:09:26.055283 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-84bc9db6cc-2f2dh" Oct 03 13:09:26 crc kubenswrapper[4578]: I1003 13:09:26.081127 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6769b867d9-dt8s5" Oct 03 13:09:26 crc kubenswrapper[4578]: I1003 13:09:26.109019 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7f55849f88-92tt9" Oct 03 13:09:26 crc kubenswrapper[4578]: I1003 13:09:26.150503 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6fd6854b49-kgt86" Oct 03 13:09:26 crc kubenswrapper[4578]: I1003 13:09:26.176430 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-5c468bf4d4-cn6bd" Oct 03 13:09:26 crc kubenswrapper[4578]: I1003 13:09:26.228723 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-555c7456bd-p4chq" Oct 03 13:09:26 crc kubenswrapper[4578]: I1003 13:09:26.283530 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-59d6cfdf45-b28kv" Oct 03 13:09:26 crc kubenswrapper[4578]: I1003 13:09:26.412020 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-7d8bb7f44c-ql5zc" Oct 03 13:09:26 crc kubenswrapper[4578]: I1003 13:09:26.568704 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-kzr6z" Oct 03 13:09:26 crc kubenswrapper[4578]: I1003 13:09:26.606532 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5db5cf686f-rpzgw" Oct 03 13:09:26 crc kubenswrapper[4578]: I1003 13:09:26.670823 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-6ft9g" Oct 03 13:09:26 crc kubenswrapper[4578]: I1003 13:09:26.672967 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-688db7b6c7-6ft9g" Oct 03 13:09:26 crc kubenswrapper[4578]: I1003 13:09:26.707819 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-fcd7d9895-dcr4b" Oct 03 13:09:27 crc kubenswrapper[4578]: I1003 13:09:27.430824 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5fbf469cd7-d6mqh" Oct 03 13:09:33 crc kubenswrapper[4578]: I1003 13:09:33.206701 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-r2zgb" event={"ID":"5b7d3469-bfd4-47c8-a82c-3e00721b2102","Type":"ContainerStarted","Data":"9cd1e91db63e8a41317711c48109477645dd9a72731483cb4c55f8adf94a4410"} Oct 03 13:09:33 crc kubenswrapper[4578]: I1003 13:09:33.207183 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-r2zgb" Oct 03 13:09:33 crc kubenswrapper[4578]: I1003 13:09:33.230425 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-r2zgb" podStartSLOduration=4.361283264 podStartE2EDuration="48.230404035s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:48.652937159 +0000 UTC m=+1064.451409343" lastFinishedPulling="2025-10-03 13:09:32.52205793 +0000 UTC m=+1108.320530114" observedRunningTime="2025-10-03 13:09:33.223441825 +0000 UTC m=+1109.021914019" watchObservedRunningTime="2025-10-03 13:09:33.230404035 +0000 UTC m=+1109.028876239" Oct 03 13:09:35 crc kubenswrapper[4578]: I1003 13:09:35.092066 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:09:35 crc kubenswrapper[4578]: I1003 13:09:35.092696 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:09:35 crc kubenswrapper[4578]: I1003 13:09:35.092899 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 13:09:35 crc kubenswrapper[4578]: I1003 13:09:35.093760 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"415155e4360685a3545a1494ed131d83906f75047c2e61272798309c3d089d8b"} pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:09:35 crc kubenswrapper[4578]: I1003 13:09:35.094000 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" containerID="cri-o://415155e4360685a3545a1494ed131d83906f75047c2e61272798309c3d089d8b" gracePeriod=600 Oct 03 13:09:35 crc kubenswrapper[4578]: I1003 13:09:35.691037 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-599898f689-vjh6c" Oct 03 13:09:36 crc kubenswrapper[4578]: I1003 13:09:36.234037 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-q5cs2" event={"ID":"86ba4ceb-a98b-473f-a133-c20b0d95ca04","Type":"ContainerStarted","Data":"8f4130acc5d47197392d904694149b44f856191de41c16890545e1825272eda9"} Oct 03 13:09:36 crc kubenswrapper[4578]: I1003 13:09:36.234357 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-q5cs2" Oct 03 13:09:36 crc kubenswrapper[4578]: I1003 13:09:36.237137 4578 generic.go:334] "Generic (PLEG): container finished" podID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerID="415155e4360685a3545a1494ed131d83906f75047c2e61272798309c3d089d8b" exitCode=0 Oct 03 13:09:36 crc kubenswrapper[4578]: I1003 13:09:36.237183 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerDied","Data":"415155e4360685a3545a1494ed131d83906f75047c2e61272798309c3d089d8b"} Oct 03 13:09:36 crc kubenswrapper[4578]: I1003 13:09:36.237225 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"ad105f3cc883a1c5d61b8c9107a9b7139beb580cccd8fece952ff237bc9b856e"} Oct 03 13:09:36 crc kubenswrapper[4578]: I1003 13:09:36.237253 4578 scope.go:117] "RemoveContainer" containerID="942bca54ab772aabe22d68921c64f6de2fc3db7407fb9842d79fce2fa9613a96" Oct 03 13:09:36 crc kubenswrapper[4578]: I1003 13:09:36.240420 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-s54bp" event={"ID":"249acba6-7e64-4f9f-8f60-22740371de34","Type":"ContainerStarted","Data":"f67d7b3731c4c03a81c74b493050b2aa0f8102ef6366efab0eb955745f0d6ea2"} Oct 03 13:09:36 crc kubenswrapper[4578]: I1003 13:09:36.240607 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-s54bp" Oct 03 13:09:36 crc kubenswrapper[4578]: I1003 13:09:36.243253 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" event={"ID":"31250d8b-75f5-4145-b8ed-1ec79bebe936","Type":"ContainerStarted","Data":"0338dfabc11e897e8ced02153774e51a30bdcfbc4415be765b4995a75739043a"} Oct 03 13:09:36 crc kubenswrapper[4578]: I1003 13:09:36.243467 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" Oct 03 13:09:36 crc kubenswrapper[4578]: I1003 13:09:36.255406 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-q5cs2" podStartSLOduration=3.145277086 podStartE2EDuration="51.25538815s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:47.38069021 +0000 UTC m=+1063.179162394" lastFinishedPulling="2025-10-03 13:09:35.490801274 +0000 UTC m=+1111.289273458" observedRunningTime="2025-10-03 13:09:36.255324718 +0000 UTC m=+1112.053796912" watchObservedRunningTime="2025-10-03 13:09:36.25538815 +0000 UTC m=+1112.053860324" Oct 03 13:09:36 crc kubenswrapper[4578]: I1003 13:09:36.276086 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-s54bp" podStartSLOduration=3.572929239 podStartE2EDuration="51.276069732s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:47.688894729 +0000 UTC m=+1063.487366913" lastFinishedPulling="2025-10-03 13:09:35.392035222 +0000 UTC m=+1111.190507406" observedRunningTime="2025-10-03 13:09:36.272564922 +0000 UTC m=+1112.071037106" watchObservedRunningTime="2025-10-03 13:09:36.276069732 +0000 UTC m=+1112.074541916" Oct 03 13:09:36 crc kubenswrapper[4578]: I1003 13:09:36.321687 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" podStartSLOduration=4.47511333 podStartE2EDuration="51.321661928s" podCreationTimestamp="2025-10-03 13:08:45 +0000 UTC" firstStartedPulling="2025-10-03 13:08:48.644459862 +0000 UTC m=+1064.442932046" lastFinishedPulling="2025-10-03 13:09:35.49100846 +0000 UTC m=+1111.289480644" observedRunningTime="2025-10-03 13:09:36.320013726 +0000 UTC m=+1112.118485910" watchObservedRunningTime="2025-10-03 13:09:36.321661928 +0000 UTC m=+1112.120134112" Oct 03 13:09:45 crc kubenswrapper[4578]: I1003 13:09:45.670979 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-846dff85b5-q5cs2" Oct 03 13:09:46 crc kubenswrapper[4578]: I1003 13:09:46.190313 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-6574bf987d-s54bp" Oct 03 13:09:46 crc kubenswrapper[4578]: I1003 13:09:46.364506 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-r2zgb" Oct 03 13:09:47 crc kubenswrapper[4578]: I1003 13:09:47.838674 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f64c4d6789s9md" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.378920 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-745tx"] Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.381023 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-745tx" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.385967 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.386068 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.386336 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.392288 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-2zccp" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.398460 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-745tx"] Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.461387 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-5lpwj"] Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.462590 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-5lpwj" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.465564 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.488483 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-5lpwj"] Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.488757 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz896\" (UniqueName: \"kubernetes.io/projected/d175b8ea-f1f6-4f08-a577-e596d7024d20-kube-api-access-tz896\") pod \"dnsmasq-dns-675f4bcbfc-745tx\" (UID: \"d175b8ea-f1f6-4f08-a577-e596d7024d20\") " pod="openstack/dnsmasq-dns-675f4bcbfc-745tx" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.488859 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d175b8ea-f1f6-4f08-a577-e596d7024d20-config\") pod \"dnsmasq-dns-675f4bcbfc-745tx\" (UID: \"d175b8ea-f1f6-4f08-a577-e596d7024d20\") " pod="openstack/dnsmasq-dns-675f4bcbfc-745tx" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.590482 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18730769-2ef4-482b-8020-7be7c37f5668-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-5lpwj\" (UID: \"18730769-2ef4-482b-8020-7be7c37f5668\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5lpwj" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.591163 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18730769-2ef4-482b-8020-7be7c37f5668-config\") pod \"dnsmasq-dns-78dd6ddcc-5lpwj\" (UID: \"18730769-2ef4-482b-8020-7be7c37f5668\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5lpwj" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.591322 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d175b8ea-f1f6-4f08-a577-e596d7024d20-config\") pod \"dnsmasq-dns-675f4bcbfc-745tx\" (UID: \"d175b8ea-f1f6-4f08-a577-e596d7024d20\") " pod="openstack/dnsmasq-dns-675f4bcbfc-745tx" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.591474 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz896\" (UniqueName: \"kubernetes.io/projected/d175b8ea-f1f6-4f08-a577-e596d7024d20-kube-api-access-tz896\") pod \"dnsmasq-dns-675f4bcbfc-745tx\" (UID: \"d175b8ea-f1f6-4f08-a577-e596d7024d20\") " pod="openstack/dnsmasq-dns-675f4bcbfc-745tx" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.591587 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrmvg\" (UniqueName: \"kubernetes.io/projected/18730769-2ef4-482b-8020-7be7c37f5668-kube-api-access-wrmvg\") pod \"dnsmasq-dns-78dd6ddcc-5lpwj\" (UID: \"18730769-2ef4-482b-8020-7be7c37f5668\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5lpwj" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.593525 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d175b8ea-f1f6-4f08-a577-e596d7024d20-config\") pod \"dnsmasq-dns-675f4bcbfc-745tx\" (UID: \"d175b8ea-f1f6-4f08-a577-e596d7024d20\") " pod="openstack/dnsmasq-dns-675f4bcbfc-745tx" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.617783 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz896\" (UniqueName: \"kubernetes.io/projected/d175b8ea-f1f6-4f08-a577-e596d7024d20-kube-api-access-tz896\") pod \"dnsmasq-dns-675f4bcbfc-745tx\" (UID: \"d175b8ea-f1f6-4f08-a577-e596d7024d20\") " pod="openstack/dnsmasq-dns-675f4bcbfc-745tx" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.693131 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18730769-2ef4-482b-8020-7be7c37f5668-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-5lpwj\" (UID: \"18730769-2ef4-482b-8020-7be7c37f5668\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5lpwj" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.693174 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18730769-2ef4-482b-8020-7be7c37f5668-config\") pod \"dnsmasq-dns-78dd6ddcc-5lpwj\" (UID: \"18730769-2ef4-482b-8020-7be7c37f5668\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5lpwj" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.693255 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrmvg\" (UniqueName: \"kubernetes.io/projected/18730769-2ef4-482b-8020-7be7c37f5668-kube-api-access-wrmvg\") pod \"dnsmasq-dns-78dd6ddcc-5lpwj\" (UID: \"18730769-2ef4-482b-8020-7be7c37f5668\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5lpwj" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.693947 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18730769-2ef4-482b-8020-7be7c37f5668-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-5lpwj\" (UID: \"18730769-2ef4-482b-8020-7be7c37f5668\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5lpwj" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.694428 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18730769-2ef4-482b-8020-7be7c37f5668-config\") pod \"dnsmasq-dns-78dd6ddcc-5lpwj\" (UID: \"18730769-2ef4-482b-8020-7be7c37f5668\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5lpwj" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.702688 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-745tx" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.721476 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrmvg\" (UniqueName: \"kubernetes.io/projected/18730769-2ef4-482b-8020-7be7c37f5668-kube-api-access-wrmvg\") pod \"dnsmasq-dns-78dd6ddcc-5lpwj\" (UID: \"18730769-2ef4-482b-8020-7be7c37f5668\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5lpwj" Oct 03 13:10:05 crc kubenswrapper[4578]: I1003 13:10:05.782657 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-5lpwj" Oct 03 13:10:06 crc kubenswrapper[4578]: I1003 13:10:06.248664 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-745tx"] Oct 03 13:10:06 crc kubenswrapper[4578]: I1003 13:10:06.321334 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-5lpwj"] Oct 03 13:10:06 crc kubenswrapper[4578]: W1003 13:10:06.324345 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18730769_2ef4_482b_8020_7be7c37f5668.slice/crio-a8f385ac2a77222a09bc46f90230ceebcc75e202daca65a8f77e7396d90b4b77 WatchSource:0}: Error finding container a8f385ac2a77222a09bc46f90230ceebcc75e202daca65a8f77e7396d90b4b77: Status 404 returned error can't find the container with id a8f385ac2a77222a09bc46f90230ceebcc75e202daca65a8f77e7396d90b4b77 Oct 03 13:10:06 crc kubenswrapper[4578]: I1003 13:10:06.467330 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-5lpwj" event={"ID":"18730769-2ef4-482b-8020-7be7c37f5668","Type":"ContainerStarted","Data":"a8f385ac2a77222a09bc46f90230ceebcc75e202daca65a8f77e7396d90b4b77"} Oct 03 13:10:06 crc kubenswrapper[4578]: I1003 13:10:06.475870 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-745tx" event={"ID":"d175b8ea-f1f6-4f08-a577-e596d7024d20","Type":"ContainerStarted","Data":"5403336d78f97b6de5b7d772387de3b97071a91afc48b9974f8069023437e6d8"} Oct 03 13:10:08 crc kubenswrapper[4578]: I1003 13:10:08.450537 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-745tx"] Oct 03 13:10:08 crc kubenswrapper[4578]: I1003 13:10:08.517796 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7xdrm"] Oct 03 13:10:08 crc kubenswrapper[4578]: I1003 13:10:08.532607 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7xdrm"] Oct 03 13:10:08 crc kubenswrapper[4578]: I1003 13:10:08.532792 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" Oct 03 13:10:08 crc kubenswrapper[4578]: I1003 13:10:08.663507 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a430efe-d234-4742-af6c-40a34b16daf6-dns-svc\") pod \"dnsmasq-dns-666b6646f7-7xdrm\" (UID: \"2a430efe-d234-4742-af6c-40a34b16daf6\") " pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" Oct 03 13:10:08 crc kubenswrapper[4578]: I1003 13:10:08.663574 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a430efe-d234-4742-af6c-40a34b16daf6-config\") pod \"dnsmasq-dns-666b6646f7-7xdrm\" (UID: \"2a430efe-d234-4742-af6c-40a34b16daf6\") " pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" Oct 03 13:10:08 crc kubenswrapper[4578]: I1003 13:10:08.663603 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkqt6\" (UniqueName: \"kubernetes.io/projected/2a430efe-d234-4742-af6c-40a34b16daf6-kube-api-access-jkqt6\") pod \"dnsmasq-dns-666b6646f7-7xdrm\" (UID: \"2a430efe-d234-4742-af6c-40a34b16daf6\") " pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" Oct 03 13:10:08 crc kubenswrapper[4578]: I1003 13:10:08.764656 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a430efe-d234-4742-af6c-40a34b16daf6-dns-svc\") pod \"dnsmasq-dns-666b6646f7-7xdrm\" (UID: \"2a430efe-d234-4742-af6c-40a34b16daf6\") " pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" Oct 03 13:10:08 crc kubenswrapper[4578]: I1003 13:10:08.764731 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a430efe-d234-4742-af6c-40a34b16daf6-config\") pod \"dnsmasq-dns-666b6646f7-7xdrm\" (UID: \"2a430efe-d234-4742-af6c-40a34b16daf6\") " pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" Oct 03 13:10:08 crc kubenswrapper[4578]: I1003 13:10:08.764760 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkqt6\" (UniqueName: \"kubernetes.io/projected/2a430efe-d234-4742-af6c-40a34b16daf6-kube-api-access-jkqt6\") pod \"dnsmasq-dns-666b6646f7-7xdrm\" (UID: \"2a430efe-d234-4742-af6c-40a34b16daf6\") " pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" Oct 03 13:10:08 crc kubenswrapper[4578]: I1003 13:10:08.766135 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a430efe-d234-4742-af6c-40a34b16daf6-config\") pod \"dnsmasq-dns-666b6646f7-7xdrm\" (UID: \"2a430efe-d234-4742-af6c-40a34b16daf6\") " pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" Oct 03 13:10:08 crc kubenswrapper[4578]: I1003 13:10:08.766179 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a430efe-d234-4742-af6c-40a34b16daf6-dns-svc\") pod \"dnsmasq-dns-666b6646f7-7xdrm\" (UID: \"2a430efe-d234-4742-af6c-40a34b16daf6\") " pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" Oct 03 13:10:08 crc kubenswrapper[4578]: I1003 13:10:08.802607 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkqt6\" (UniqueName: \"kubernetes.io/projected/2a430efe-d234-4742-af6c-40a34b16daf6-kube-api-access-jkqt6\") pod \"dnsmasq-dns-666b6646f7-7xdrm\" (UID: \"2a430efe-d234-4742-af6c-40a34b16daf6\") " pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" Oct 03 13:10:08 crc kubenswrapper[4578]: I1003 13:10:08.862296 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-5lpwj"] Oct 03 13:10:08 crc kubenswrapper[4578]: I1003 13:10:08.881182 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" Oct 03 13:10:08 crc kubenswrapper[4578]: I1003 13:10:08.897553 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-z97ct"] Oct 03 13:10:08 crc kubenswrapper[4578]: I1003 13:10:08.898714 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" Oct 03 13:10:08 crc kubenswrapper[4578]: I1003 13:10:08.912325 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-z97ct"] Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.075168 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq8jc\" (UniqueName: \"kubernetes.io/projected/dc1c7337-6ef2-409f-b307-6f1cf60ee599-kube-api-access-kq8jc\") pod \"dnsmasq-dns-57d769cc4f-z97ct\" (UID: \"dc1c7337-6ef2-409f-b307-6f1cf60ee599\") " pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.075266 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc1c7337-6ef2-409f-b307-6f1cf60ee599-config\") pod \"dnsmasq-dns-57d769cc4f-z97ct\" (UID: \"dc1c7337-6ef2-409f-b307-6f1cf60ee599\") " pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.075286 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc1c7337-6ef2-409f-b307-6f1cf60ee599-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-z97ct\" (UID: \"dc1c7337-6ef2-409f-b307-6f1cf60ee599\") " pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.187615 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq8jc\" (UniqueName: \"kubernetes.io/projected/dc1c7337-6ef2-409f-b307-6f1cf60ee599-kube-api-access-kq8jc\") pod \"dnsmasq-dns-57d769cc4f-z97ct\" (UID: \"dc1c7337-6ef2-409f-b307-6f1cf60ee599\") " pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.187699 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc1c7337-6ef2-409f-b307-6f1cf60ee599-config\") pod \"dnsmasq-dns-57d769cc4f-z97ct\" (UID: \"dc1c7337-6ef2-409f-b307-6f1cf60ee599\") " pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.187720 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc1c7337-6ef2-409f-b307-6f1cf60ee599-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-z97ct\" (UID: \"dc1c7337-6ef2-409f-b307-6f1cf60ee599\") " pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.188683 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc1c7337-6ef2-409f-b307-6f1cf60ee599-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-z97ct\" (UID: \"dc1c7337-6ef2-409f-b307-6f1cf60ee599\") " pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.189518 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc1c7337-6ef2-409f-b307-6f1cf60ee599-config\") pod \"dnsmasq-dns-57d769cc4f-z97ct\" (UID: \"dc1c7337-6ef2-409f-b307-6f1cf60ee599\") " pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.252423 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq8jc\" (UniqueName: \"kubernetes.io/projected/dc1c7337-6ef2-409f-b307-6f1cf60ee599-kube-api-access-kq8jc\") pod \"dnsmasq-dns-57d769cc4f-z97ct\" (UID: \"dc1c7337-6ef2-409f-b307-6f1cf60ee599\") " pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.282513 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.601898 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7xdrm"] Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.692068 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.698115 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.713137 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.713240 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.713367 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-nwbq9" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.713422 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.713481 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.713573 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.713620 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.722724 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.807602 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.807705 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.807820 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c196bfb1-793e-4e04-9602-db06886385ad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.807919 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh8vr\" (UniqueName: \"kubernetes.io/projected/c196bfb1-793e-4e04-9602-db06886385ad-kube-api-access-bh8vr\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.807948 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.807993 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.808052 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.808204 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c196bfb1-793e-4e04-9602-db06886385ad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.808277 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c196bfb1-793e-4e04-9602-db06886385ad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.808309 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c196bfb1-793e-4e04-9602-db06886385ad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.808371 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c196bfb1-793e-4e04-9602-db06886385ad-config-data\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.914413 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c196bfb1-793e-4e04-9602-db06886385ad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.914472 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c196bfb1-793e-4e04-9602-db06886385ad-config-data\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.914501 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.914521 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.914552 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c196bfb1-793e-4e04-9602-db06886385ad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.914574 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.914592 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh8vr\" (UniqueName: \"kubernetes.io/projected/c196bfb1-793e-4e04-9602-db06886385ad-kube-api-access-bh8vr\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.914643 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.914682 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.914714 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c196bfb1-793e-4e04-9602-db06886385ad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.914751 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c196bfb1-793e-4e04-9602-db06886385ad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.915320 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.916541 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c196bfb1-793e-4e04-9602-db06886385ad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.916604 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c196bfb1-793e-4e04-9602-db06886385ad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.916860 4578 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.917345 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c196bfb1-793e-4e04-9602-db06886385ad-config-data\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.918000 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.928479 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.933896 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c196bfb1-793e-4e04-9602-db06886385ad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.948655 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.963291 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c196bfb1-793e-4e04-9602-db06886385ad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.993196 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:09 crc kubenswrapper[4578]: I1003 13:10:09.994511 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh8vr\" (UniqueName: \"kubernetes.io/projected/c196bfb1-793e-4e04-9602-db06886385ad-kube-api-access-bh8vr\") pod \"rabbitmq-server-0\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " pod="openstack/rabbitmq-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.054045 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.070184 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-z97ct"] Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.116334 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.117590 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.120935 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.121139 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.121249 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-wxlrl" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.121347 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.121477 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.121616 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.121718 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.137976 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.222618 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.222669 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.222700 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.222724 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.222743 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/79fcccfa-6df4-41f2-965b-357e5a7984ab-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.222770 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9zzn\" (UniqueName: \"kubernetes.io/projected/79fcccfa-6df4-41f2-965b-357e5a7984ab-kube-api-access-z9zzn\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.222797 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/79fcccfa-6df4-41f2-965b-357e5a7984ab-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.222816 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.222833 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/79fcccfa-6df4-41f2-965b-357e5a7984ab-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.222860 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/79fcccfa-6df4-41f2-965b-357e5a7984ab-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.222876 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79fcccfa-6df4-41f2-965b-357e5a7984ab-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.324562 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/79fcccfa-6df4-41f2-965b-357e5a7984ab-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.324618 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79fcccfa-6df4-41f2-965b-357e5a7984ab-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.324656 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/79fcccfa-6df4-41f2-965b-357e5a7984ab-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.324701 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.324718 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.324743 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.324764 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.324782 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/79fcccfa-6df4-41f2-965b-357e5a7984ab-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.324807 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9zzn\" (UniqueName: \"kubernetes.io/projected/79fcccfa-6df4-41f2-965b-357e5a7984ab-kube-api-access-z9zzn\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.324833 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/79fcccfa-6df4-41f2-965b-357e5a7984ab-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.324849 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.325286 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.325545 4578 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.326909 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/79fcccfa-6df4-41f2-965b-357e5a7984ab-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.327439 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79fcccfa-6df4-41f2-965b-357e5a7984ab-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.329168 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.330105 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/79fcccfa-6df4-41f2-965b-357e5a7984ab-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.332349 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/79fcccfa-6df4-41f2-965b-357e5a7984ab-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.332588 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.333220 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/79fcccfa-6df4-41f2-965b-357e5a7984ab-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.334771 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.348058 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9zzn\" (UniqueName: \"kubernetes.io/projected/79fcccfa-6df4-41f2-965b-357e5a7984ab-kube-api-access-z9zzn\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.354916 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.459512 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.540611 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" event={"ID":"2a430efe-d234-4742-af6c-40a34b16daf6","Type":"ContainerStarted","Data":"10fd72ed7c3aa3dffca5c9d0958c1542d926153a8597b7777b97794b7478750b"} Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.543465 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" event={"ID":"dc1c7337-6ef2-409f-b307-6f1cf60ee599","Type":"ContainerStarted","Data":"59c8020257a7ae607289ed85e0b110031b3f1c8e5544ecc7598c3ce749a7b790"} Oct 03 13:10:10 crc kubenswrapper[4578]: I1003 13:10:10.618052 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 13:10:11 crc kubenswrapper[4578]: I1003 13:10:11.064252 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 13:10:11 crc kubenswrapper[4578]: I1003 13:10:11.647944 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c196bfb1-793e-4e04-9602-db06886385ad","Type":"ContainerStarted","Data":"b26db1e2f552b04b5d2c308687c6df6b76bbdcb0e3ef58559e8711050d5b6547"} Oct 03 13:10:11 crc kubenswrapper[4578]: I1003 13:10:11.710122 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"79fcccfa-6df4-41f2-965b-357e5a7984ab","Type":"ContainerStarted","Data":"1c17bc81fb7e6a865cc8a53c6027ca55b08892125ecd4876ede2e6717a1c23c3"} Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.062463 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.064565 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.073294 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.073474 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-c6pt5" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.073578 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.074219 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.074454 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.086561 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.095116 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.110216 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.113677 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.119847 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.120185 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.120386 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-8qgwv" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.123024 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.125115 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zctrm\" (UniqueName: \"kubernetes.io/projected/62242ee6-b3ff-4771-a989-4d7bd1d243d3-kube-api-access-zctrm\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.125162 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.125192 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62242ee6-b3ff-4771-a989-4d7bd1d243d3-operator-scripts\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.125227 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/62242ee6-b3ff-4771-a989-4d7bd1d243d3-config-data-default\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.125257 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62242ee6-b3ff-4771-a989-4d7bd1d243d3-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.125301 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/62242ee6-b3ff-4771-a989-4d7bd1d243d3-secrets\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.125324 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/62242ee6-b3ff-4771-a989-4d7bd1d243d3-config-data-generated\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.125345 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/62242ee6-b3ff-4771-a989-4d7bd1d243d3-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.125362 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/62242ee6-b3ff-4771-a989-4d7bd1d243d3-kolla-config\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.164403 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.231983 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/62242ee6-b3ff-4771-a989-4d7bd1d243d3-config-data-generated\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.232051 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/62242ee6-b3ff-4771-a989-4d7bd1d243d3-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.232075 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/62242ee6-b3ff-4771-a989-4d7bd1d243d3-kolla-config\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.232127 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zctrm\" (UniqueName: \"kubernetes.io/projected/62242ee6-b3ff-4771-a989-4d7bd1d243d3-kube-api-access-zctrm\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.232160 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.232194 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62242ee6-b3ff-4771-a989-4d7bd1d243d3-operator-scripts\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.233277 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/62242ee6-b3ff-4771-a989-4d7bd1d243d3-config-data-default\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.233349 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62242ee6-b3ff-4771-a989-4d7bd1d243d3-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.233387 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/62242ee6-b3ff-4771-a989-4d7bd1d243d3-secrets\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.247980 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/62242ee6-b3ff-4771-a989-4d7bd1d243d3-config-data-generated\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.249727 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/62242ee6-b3ff-4771-a989-4d7bd1d243d3-config-data-default\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.249868 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62242ee6-b3ff-4771-a989-4d7bd1d243d3-operator-scripts\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.250108 4578 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.251065 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/62242ee6-b3ff-4771-a989-4d7bd1d243d3-secrets\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.256436 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/62242ee6-b3ff-4771-a989-4d7bd1d243d3-kolla-config\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.257025 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62242ee6-b3ff-4771-a989-4d7bd1d243d3-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.264673 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/62242ee6-b3ff-4771-a989-4d7bd1d243d3-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.288868 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zctrm\" (UniqueName: \"kubernetes.io/projected/62242ee6-b3ff-4771-a989-4d7bd1d243d3-kube-api-access-zctrm\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.297252 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"62242ee6-b3ff-4771-a989-4d7bd1d243d3\") " pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.335091 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/32caee81-c7c2-45ee-bb87-2e6003d6115d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.335152 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/32caee81-c7c2-45ee-bb87-2e6003d6115d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.335182 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/32caee81-c7c2-45ee-bb87-2e6003d6115d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.335208 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32caee81-c7c2-45ee-bb87-2e6003d6115d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.335243 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/32caee81-c7c2-45ee-bb87-2e6003d6115d-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.335269 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/32caee81-c7c2-45ee-bb87-2e6003d6115d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.335287 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.335309 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32caee81-c7c2-45ee-bb87-2e6003d6115d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.335326 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6jmm\" (UniqueName: \"kubernetes.io/projected/32caee81-c7c2-45ee-bb87-2e6003d6115d-kube-api-access-h6jmm\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.385535 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.436588 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/32caee81-c7c2-45ee-bb87-2e6003d6115d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.436708 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/32caee81-c7c2-45ee-bb87-2e6003d6115d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.436729 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/32caee81-c7c2-45ee-bb87-2e6003d6115d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.436758 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32caee81-c7c2-45ee-bb87-2e6003d6115d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.436791 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/32caee81-c7c2-45ee-bb87-2e6003d6115d-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.436820 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/32caee81-c7c2-45ee-bb87-2e6003d6115d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.436842 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.436864 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32caee81-c7c2-45ee-bb87-2e6003d6115d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.436885 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6jmm\" (UniqueName: \"kubernetes.io/projected/32caee81-c7c2-45ee-bb87-2e6003d6115d-kube-api-access-h6jmm\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.437141 4578 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.439004 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32caee81-c7c2-45ee-bb87-2e6003d6115d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.439529 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/32caee81-c7c2-45ee-bb87-2e6003d6115d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.439887 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/32caee81-c7c2-45ee-bb87-2e6003d6115d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.449528 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/32caee81-c7c2-45ee-bb87-2e6003d6115d-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.452713 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32caee81-c7c2-45ee-bb87-2e6003d6115d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.456534 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/32caee81-c7c2-45ee-bb87-2e6003d6115d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.457863 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/32caee81-c7c2-45ee-bb87-2e6003d6115d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.513727 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.522075 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6jmm\" (UniqueName: \"kubernetes.io/projected/32caee81-c7c2-45ee-bb87-2e6003d6115d-kube-api-access-h6jmm\") pod \"openstack-cell1-galera-0\" (UID: \"32caee81-c7c2-45ee-bb87-2e6003d6115d\") " pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.691374 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.693012 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.706060 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-kk787" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.706337 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.706439 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.712717 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.739471 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.845831 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8v5q\" (UniqueName: \"kubernetes.io/projected/409ed574-cc6d-49e4-b1b1-4ec50dfe171e-kube-api-access-v8v5q\") pod \"memcached-0\" (UID: \"409ed574-cc6d-49e4-b1b1-4ec50dfe171e\") " pod="openstack/memcached-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.846182 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/409ed574-cc6d-49e4-b1b1-4ec50dfe171e-config-data\") pod \"memcached-0\" (UID: \"409ed574-cc6d-49e4-b1b1-4ec50dfe171e\") " pod="openstack/memcached-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.846229 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/409ed574-cc6d-49e4-b1b1-4ec50dfe171e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"409ed574-cc6d-49e4-b1b1-4ec50dfe171e\") " pod="openstack/memcached-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.846253 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/409ed574-cc6d-49e4-b1b1-4ec50dfe171e-kolla-config\") pod \"memcached-0\" (UID: \"409ed574-cc6d-49e4-b1b1-4ec50dfe171e\") " pod="openstack/memcached-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.846276 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/409ed574-cc6d-49e4-b1b1-4ec50dfe171e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"409ed574-cc6d-49e4-b1b1-4ec50dfe171e\") " pod="openstack/memcached-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.948195 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/409ed574-cc6d-49e4-b1b1-4ec50dfe171e-kolla-config\") pod \"memcached-0\" (UID: \"409ed574-cc6d-49e4-b1b1-4ec50dfe171e\") " pod="openstack/memcached-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.948257 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/409ed574-cc6d-49e4-b1b1-4ec50dfe171e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"409ed574-cc6d-49e4-b1b1-4ec50dfe171e\") " pod="openstack/memcached-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.948331 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8v5q\" (UniqueName: \"kubernetes.io/projected/409ed574-cc6d-49e4-b1b1-4ec50dfe171e-kube-api-access-v8v5q\") pod \"memcached-0\" (UID: \"409ed574-cc6d-49e4-b1b1-4ec50dfe171e\") " pod="openstack/memcached-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.948364 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/409ed574-cc6d-49e4-b1b1-4ec50dfe171e-config-data\") pod \"memcached-0\" (UID: \"409ed574-cc6d-49e4-b1b1-4ec50dfe171e\") " pod="openstack/memcached-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.948405 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/409ed574-cc6d-49e4-b1b1-4ec50dfe171e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"409ed574-cc6d-49e4-b1b1-4ec50dfe171e\") " pod="openstack/memcached-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.950473 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/409ed574-cc6d-49e4-b1b1-4ec50dfe171e-kolla-config\") pod \"memcached-0\" (UID: \"409ed574-cc6d-49e4-b1b1-4ec50dfe171e\") " pod="openstack/memcached-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.950547 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/409ed574-cc6d-49e4-b1b1-4ec50dfe171e-config-data\") pod \"memcached-0\" (UID: \"409ed574-cc6d-49e4-b1b1-4ec50dfe171e\") " pod="openstack/memcached-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.956468 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/409ed574-cc6d-49e4-b1b1-4ec50dfe171e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"409ed574-cc6d-49e4-b1b1-4ec50dfe171e\") " pod="openstack/memcached-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.961055 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/409ed574-cc6d-49e4-b1b1-4ec50dfe171e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"409ed574-cc6d-49e4-b1b1-4ec50dfe171e\") " pod="openstack/memcached-0" Oct 03 13:10:13 crc kubenswrapper[4578]: I1003 13:10:13.976059 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8v5q\" (UniqueName: \"kubernetes.io/projected/409ed574-cc6d-49e4-b1b1-4ec50dfe171e-kube-api-access-v8v5q\") pod \"memcached-0\" (UID: \"409ed574-cc6d-49e4-b1b1-4ec50dfe171e\") " pod="openstack/memcached-0" Oct 03 13:10:14 crc kubenswrapper[4578]: I1003 13:10:14.029768 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 03 13:10:14 crc kubenswrapper[4578]: I1003 13:10:14.325932 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 03 13:10:14 crc kubenswrapper[4578]: I1003 13:10:14.637754 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 03 13:10:14 crc kubenswrapper[4578]: I1003 13:10:14.860319 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"32caee81-c7c2-45ee-bb87-2e6003d6115d","Type":"ContainerStarted","Data":"d97f95964c6c0ec3565258b75a58d844ca887ca9b67756e2e479fcd877051c7f"} Oct 03 13:10:14 crc kubenswrapper[4578]: I1003 13:10:14.960505 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"62242ee6-b3ff-4771-a989-4d7bd1d243d3","Type":"ContainerStarted","Data":"c25206f6ccd710b27b082e5d7ff3a693c8d8ea81f3b6317be9dbb63d20be4f74"} Oct 03 13:10:15 crc kubenswrapper[4578]: I1003 13:10:15.126803 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 03 13:10:15 crc kubenswrapper[4578]: W1003 13:10:15.143454 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod409ed574_cc6d_49e4_b1b1_4ec50dfe171e.slice/crio-4fb33a3675748281507d9c5af26de190795a3cfd1bf6466775d6298a19d8c5f9 WatchSource:0}: Error finding container 4fb33a3675748281507d9c5af26de190795a3cfd1bf6466775d6298a19d8c5f9: Status 404 returned error can't find the container with id 4fb33a3675748281507d9c5af26de190795a3cfd1bf6466775d6298a19d8c5f9 Oct 03 13:10:16 crc kubenswrapper[4578]: I1003 13:10:15.986842 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"409ed574-cc6d-49e4-b1b1-4ec50dfe171e","Type":"ContainerStarted","Data":"4fb33a3675748281507d9c5af26de190795a3cfd1bf6466775d6298a19d8c5f9"} Oct 03 13:10:16 crc kubenswrapper[4578]: I1003 13:10:16.510884 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:10:16 crc kubenswrapper[4578]: I1003 13:10:16.512106 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 13:10:16 crc kubenswrapper[4578]: I1003 13:10:16.514923 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-fc2x6" Oct 03 13:10:16 crc kubenswrapper[4578]: I1003 13:10:16.516962 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:10:16 crc kubenswrapper[4578]: I1003 13:10:16.641323 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8hdw\" (UniqueName: \"kubernetes.io/projected/58664f3d-6768-4b3b-a96a-6c22c2c7a8c4-kube-api-access-f8hdw\") pod \"kube-state-metrics-0\" (UID: \"58664f3d-6768-4b3b-a96a-6c22c2c7a8c4\") " pod="openstack/kube-state-metrics-0" Oct 03 13:10:16 crc kubenswrapper[4578]: I1003 13:10:16.749682 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8hdw\" (UniqueName: \"kubernetes.io/projected/58664f3d-6768-4b3b-a96a-6c22c2c7a8c4-kube-api-access-f8hdw\") pod \"kube-state-metrics-0\" (UID: \"58664f3d-6768-4b3b-a96a-6c22c2c7a8c4\") " pod="openstack/kube-state-metrics-0" Oct 03 13:10:16 crc kubenswrapper[4578]: I1003 13:10:16.795330 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8hdw\" (UniqueName: \"kubernetes.io/projected/58664f3d-6768-4b3b-a96a-6c22c2c7a8c4-kube-api-access-f8hdw\") pod \"kube-state-metrics-0\" (UID: \"58664f3d-6768-4b3b-a96a-6c22c2c7a8c4\") " pod="openstack/kube-state-metrics-0" Oct 03 13:10:16 crc kubenswrapper[4578]: I1003 13:10:16.836933 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 13:10:17 crc kubenswrapper[4578]: I1003 13:10:17.511519 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:10:17 crc kubenswrapper[4578]: W1003 13:10:17.571952 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58664f3d_6768_4b3b_a96a_6c22c2c7a8c4.slice/crio-c885218c52ef95393d306068fb1041f26c7f51dc99c539fe6e65c5ba30864457 WatchSource:0}: Error finding container c885218c52ef95393d306068fb1041f26c7f51dc99c539fe6e65c5ba30864457: Status 404 returned error can't find the container with id c885218c52ef95393d306068fb1041f26c7f51dc99c539fe6e65c5ba30864457 Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.020897 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"58664f3d-6768-4b3b-a96a-6c22c2c7a8c4","Type":"ContainerStarted","Data":"c885218c52ef95393d306068fb1041f26c7f51dc99c539fe6e65c5ba30864457"} Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.496300 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-9jl4l"] Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.497525 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9jl4l"] Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.497601 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.512010 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.512209 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-8ttc5" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.518301 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-2sqbb"] Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.518942 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.520078 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.536559 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-2sqbb"] Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.700937 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-ovn-controller-tls-certs\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.700992 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-var-run-ovn\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.701093 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-var-run\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.701125 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g9ht\" (UniqueName: \"kubernetes.io/projected/14bc2fd3-f370-48d0-8711-168b0258f21f-kube-api-access-4g9ht\") pod \"ovn-controller-ovs-2sqbb\" (UID: \"14bc2fd3-f370-48d0-8711-168b0258f21f\") " pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.701140 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/14bc2fd3-f370-48d0-8711-168b0258f21f-var-lib\") pod \"ovn-controller-ovs-2sqbb\" (UID: \"14bc2fd3-f370-48d0-8711-168b0258f21f\") " pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.701203 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-scripts\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.701242 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14bc2fd3-f370-48d0-8711-168b0258f21f-scripts\") pod \"ovn-controller-ovs-2sqbb\" (UID: \"14bc2fd3-f370-48d0-8711-168b0258f21f\") " pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.701263 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-combined-ca-bundle\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.701293 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8stnc\" (UniqueName: \"kubernetes.io/projected/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-kube-api-access-8stnc\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.701325 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/14bc2fd3-f370-48d0-8711-168b0258f21f-var-log\") pod \"ovn-controller-ovs-2sqbb\" (UID: \"14bc2fd3-f370-48d0-8711-168b0258f21f\") " pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.701344 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-var-log-ovn\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.701365 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/14bc2fd3-f370-48d0-8711-168b0258f21f-etc-ovs\") pod \"ovn-controller-ovs-2sqbb\" (UID: \"14bc2fd3-f370-48d0-8711-168b0258f21f\") " pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.701379 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/14bc2fd3-f370-48d0-8711-168b0258f21f-var-run\") pod \"ovn-controller-ovs-2sqbb\" (UID: \"14bc2fd3-f370-48d0-8711-168b0258f21f\") " pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.802123 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-scripts\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.802174 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14bc2fd3-f370-48d0-8711-168b0258f21f-scripts\") pod \"ovn-controller-ovs-2sqbb\" (UID: \"14bc2fd3-f370-48d0-8711-168b0258f21f\") " pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.802190 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-combined-ca-bundle\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.802214 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8stnc\" (UniqueName: \"kubernetes.io/projected/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-kube-api-access-8stnc\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.802249 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/14bc2fd3-f370-48d0-8711-168b0258f21f-var-log\") pod \"ovn-controller-ovs-2sqbb\" (UID: \"14bc2fd3-f370-48d0-8711-168b0258f21f\") " pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.802265 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-var-log-ovn\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.802293 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/14bc2fd3-f370-48d0-8711-168b0258f21f-etc-ovs\") pod \"ovn-controller-ovs-2sqbb\" (UID: \"14bc2fd3-f370-48d0-8711-168b0258f21f\") " pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.802312 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/14bc2fd3-f370-48d0-8711-168b0258f21f-var-run\") pod \"ovn-controller-ovs-2sqbb\" (UID: \"14bc2fd3-f370-48d0-8711-168b0258f21f\") " pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.802332 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-ovn-controller-tls-certs\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.802350 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-var-run-ovn\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.802380 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-var-run\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.802406 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g9ht\" (UniqueName: \"kubernetes.io/projected/14bc2fd3-f370-48d0-8711-168b0258f21f-kube-api-access-4g9ht\") pod \"ovn-controller-ovs-2sqbb\" (UID: \"14bc2fd3-f370-48d0-8711-168b0258f21f\") " pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.802424 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/14bc2fd3-f370-48d0-8711-168b0258f21f-var-lib\") pod \"ovn-controller-ovs-2sqbb\" (UID: \"14bc2fd3-f370-48d0-8711-168b0258f21f\") " pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.802962 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/14bc2fd3-f370-48d0-8711-168b0258f21f-var-lib\") pod \"ovn-controller-ovs-2sqbb\" (UID: \"14bc2fd3-f370-48d0-8711-168b0258f21f\") " pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.806399 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14bc2fd3-f370-48d0-8711-168b0258f21f-scripts\") pod \"ovn-controller-ovs-2sqbb\" (UID: \"14bc2fd3-f370-48d0-8711-168b0258f21f\") " pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.807616 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/14bc2fd3-f370-48d0-8711-168b0258f21f-etc-ovs\") pod \"ovn-controller-ovs-2sqbb\" (UID: \"14bc2fd3-f370-48d0-8711-168b0258f21f\") " pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.807951 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-var-log-ovn\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.808024 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-var-run-ovn\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.808089 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/14bc2fd3-f370-48d0-8711-168b0258f21f-var-run\") pod \"ovn-controller-ovs-2sqbb\" (UID: \"14bc2fd3-f370-48d0-8711-168b0258f21f\") " pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.808114 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-var-run\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.808312 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/14bc2fd3-f370-48d0-8711-168b0258f21f-var-log\") pod \"ovn-controller-ovs-2sqbb\" (UID: \"14bc2fd3-f370-48d0-8711-168b0258f21f\") " pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.810031 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-scripts\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.830434 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-combined-ca-bundle\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.916911 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g9ht\" (UniqueName: \"kubernetes.io/projected/14bc2fd3-f370-48d0-8711-168b0258f21f-kube-api-access-4g9ht\") pod \"ovn-controller-ovs-2sqbb\" (UID: \"14bc2fd3-f370-48d0-8711-168b0258f21f\") " pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.917384 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-ovn-controller-tls-certs\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.919729 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:10:18 crc kubenswrapper[4578]: I1003 13:10:18.956811 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8stnc\" (UniqueName: \"kubernetes.io/projected/daabdc14-9eaa-478d-8e0f-3ab92c0568ce-kube-api-access-8stnc\") pod \"ovn-controller-9jl4l\" (UID: \"daabdc14-9eaa-478d-8e0f-3ab92c0568ce\") " pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:19 crc kubenswrapper[4578]: I1003 13:10:19.197284 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9jl4l" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.486495 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.488482 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.494028 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-86v5g" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.494269 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.494423 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.495056 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.495555 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.498147 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.652702 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ff8w\" (UniqueName: \"kubernetes.io/projected/df1fc342-37e6-4757-b1cf-81a10245997b-kube-api-access-9ff8w\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.652861 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df1fc342-37e6-4757-b1cf-81a10245997b-config\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.652895 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/df1fc342-37e6-4757-b1cf-81a10245997b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.652928 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/df1fc342-37e6-4757-b1cf-81a10245997b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.652956 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/df1fc342-37e6-4757-b1cf-81a10245997b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.652975 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df1fc342-37e6-4757-b1cf-81a10245997b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.653027 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df1fc342-37e6-4757-b1cf-81a10245997b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.653057 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.755300 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df1fc342-37e6-4757-b1cf-81a10245997b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.755360 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df1fc342-37e6-4757-b1cf-81a10245997b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.755462 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.755557 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ff8w\" (UniqueName: \"kubernetes.io/projected/df1fc342-37e6-4757-b1cf-81a10245997b-kube-api-access-9ff8w\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.755600 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df1fc342-37e6-4757-b1cf-81a10245997b-config\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.755665 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/df1fc342-37e6-4757-b1cf-81a10245997b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.755710 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/df1fc342-37e6-4757-b1cf-81a10245997b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.755744 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/df1fc342-37e6-4757-b1cf-81a10245997b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.757364 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df1fc342-37e6-4757-b1cf-81a10245997b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.757354 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df1fc342-37e6-4757-b1cf-81a10245997b-config\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.757702 4578 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.763598 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/df1fc342-37e6-4757-b1cf-81a10245997b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.766467 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/df1fc342-37e6-4757-b1cf-81a10245997b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.774364 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/df1fc342-37e6-4757-b1cf-81a10245997b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.777089 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df1fc342-37e6-4757-b1cf-81a10245997b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.798049 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ff8w\" (UniqueName: \"kubernetes.io/projected/df1fc342-37e6-4757-b1cf-81a10245997b-kube-api-access-9ff8w\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.806454 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"df1fc342-37e6-4757-b1cf-81a10245997b\") " pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:20 crc kubenswrapper[4578]: I1003 13:10:20.816003 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 03 13:10:22 crc kubenswrapper[4578]: I1003 13:10:22.906799 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 03 13:10:22 crc kubenswrapper[4578]: I1003 13:10:22.912343 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:22 crc kubenswrapper[4578]: I1003 13:10:22.915937 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 03 13:10:22 crc kubenswrapper[4578]: I1003 13:10:22.916437 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 03 13:10:22 crc kubenswrapper[4578]: I1003 13:10:22.917920 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 03 13:10:22 crc kubenswrapper[4578]: I1003 13:10:22.918104 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-z567v" Oct 03 13:10:22 crc kubenswrapper[4578]: I1003 13:10:22.924514 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 03 13:10:22 crc kubenswrapper[4578]: I1003 13:10:22.993678 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/efb87daa-8d73-444a-94ed-9e44006bec59-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:22 crc kubenswrapper[4578]: I1003 13:10:22.993727 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/efb87daa-8d73-444a-94ed-9e44006bec59-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:22 crc kubenswrapper[4578]: I1003 13:10:22.993795 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efb87daa-8d73-444a-94ed-9e44006bec59-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:22 crc kubenswrapper[4578]: I1003 13:10:22.993883 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/efb87daa-8d73-444a-94ed-9e44006bec59-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:22 crc kubenswrapper[4578]: I1003 13:10:22.994030 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:22 crc kubenswrapper[4578]: I1003 13:10:22.994068 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/efb87daa-8d73-444a-94ed-9e44006bec59-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:22 crc kubenswrapper[4578]: I1003 13:10:22.994426 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwvdn\" (UniqueName: \"kubernetes.io/projected/efb87daa-8d73-444a-94ed-9e44006bec59-kube-api-access-zwvdn\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:22 crc kubenswrapper[4578]: I1003 13:10:22.994486 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efb87daa-8d73-444a-94ed-9e44006bec59-config\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:23 crc kubenswrapper[4578]: I1003 13:10:23.096473 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:23 crc kubenswrapper[4578]: I1003 13:10:23.096525 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/efb87daa-8d73-444a-94ed-9e44006bec59-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:23 crc kubenswrapper[4578]: I1003 13:10:23.096605 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwvdn\" (UniqueName: \"kubernetes.io/projected/efb87daa-8d73-444a-94ed-9e44006bec59-kube-api-access-zwvdn\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:23 crc kubenswrapper[4578]: I1003 13:10:23.096641 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efb87daa-8d73-444a-94ed-9e44006bec59-config\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:23 crc kubenswrapper[4578]: I1003 13:10:23.096687 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/efb87daa-8d73-444a-94ed-9e44006bec59-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:23 crc kubenswrapper[4578]: I1003 13:10:23.096702 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/efb87daa-8d73-444a-94ed-9e44006bec59-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:23 crc kubenswrapper[4578]: I1003 13:10:23.096735 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efb87daa-8d73-444a-94ed-9e44006bec59-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:23 crc kubenswrapper[4578]: I1003 13:10:23.096774 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/efb87daa-8d73-444a-94ed-9e44006bec59-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:23 crc kubenswrapper[4578]: I1003 13:10:23.096866 4578 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:23 crc kubenswrapper[4578]: I1003 13:10:23.097410 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/efb87daa-8d73-444a-94ed-9e44006bec59-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:23 crc kubenswrapper[4578]: I1003 13:10:23.097956 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efb87daa-8d73-444a-94ed-9e44006bec59-config\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:23 crc kubenswrapper[4578]: I1003 13:10:23.100090 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/efb87daa-8d73-444a-94ed-9e44006bec59-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:23 crc kubenswrapper[4578]: I1003 13:10:23.102919 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efb87daa-8d73-444a-94ed-9e44006bec59-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:23 crc kubenswrapper[4578]: I1003 13:10:23.111194 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/efb87daa-8d73-444a-94ed-9e44006bec59-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:23 crc kubenswrapper[4578]: I1003 13:10:23.111218 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/efb87daa-8d73-444a-94ed-9e44006bec59-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:23 crc kubenswrapper[4578]: I1003 13:10:23.128844 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwvdn\" (UniqueName: \"kubernetes.io/projected/efb87daa-8d73-444a-94ed-9e44006bec59-kube-api-access-zwvdn\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:23 crc kubenswrapper[4578]: I1003 13:10:23.145425 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"efb87daa-8d73-444a-94ed-9e44006bec59\") " pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:23 crc kubenswrapper[4578]: I1003 13:10:23.236274 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 03 13:10:30 crc kubenswrapper[4578]: I1003 13:10:30.467809 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-2sqbb"] Oct 03 13:10:35 crc kubenswrapper[4578]: E1003 13:10:35.085956 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Oct 03 13:10:35 crc kubenswrapper[4578]: E1003 13:10:35.086594 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bh8vr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(c196bfb1-793e-4e04-9602-db06886385ad): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:10:35 crc kubenswrapper[4578]: E1003 13:10:35.088703 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="c196bfb1-793e-4e04-9602-db06886385ad" Oct 03 13:10:35 crc kubenswrapper[4578]: E1003 13:10:35.302726 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="c196bfb1-793e-4e04-9602-db06886385ad" Oct 03 13:10:37 crc kubenswrapper[4578]: E1003 13:10:37.050712 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Oct 03 13:10:37 crc kubenswrapper[4578]: E1003 13:10:37.051156 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z9zzn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(79fcccfa-6df4-41f2-965b-357e5a7984ab): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:10:37 crc kubenswrapper[4578]: E1003 13:10:37.052955 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="79fcccfa-6df4-41f2-965b-357e5a7984ab" Oct 03 13:10:37 crc kubenswrapper[4578]: E1003 13:10:37.057018 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Oct 03 13:10:37 crc kubenswrapper[4578]: E1003 13:10:37.057176 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:DB_ROOT_PASSWORD,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:DbRootPassword,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secrets,ReadOnly:true,MountPath:/var/lib/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zctrm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(62242ee6-b3ff-4771-a989-4d7bd1d243d3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:10:37 crc kubenswrapper[4578]: E1003 13:10:37.058381 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="62242ee6-b3ff-4771-a989-4d7bd1d243d3" Oct 03 13:10:37 crc kubenswrapper[4578]: E1003 13:10:37.165156 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Oct 03 13:10:37 crc kubenswrapper[4578]: E1003 13:10:37.165899 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:DB_ROOT_PASSWORD,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:DbRootPassword,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secrets,ReadOnly:true,MountPath:/var/lib/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h6jmm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(32caee81-c7c2-45ee-bb87-2e6003d6115d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:10:37 crc kubenswrapper[4578]: E1003 13:10:37.167138 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="32caee81-c7c2-45ee-bb87-2e6003d6115d" Oct 03 13:10:37 crc kubenswrapper[4578]: I1003 13:10:37.314515 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2sqbb" event={"ID":"14bc2fd3-f370-48d0-8711-168b0258f21f","Type":"ContainerStarted","Data":"24c2a9acd63b4a2e0777d11e21d60a950078bd39ab3cf1a2f63663bb6140acf7"} Oct 03 13:10:37 crc kubenswrapper[4578]: E1003 13:10:37.316240 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="32caee81-c7c2-45ee-bb87-2e6003d6115d" Oct 03 13:10:37 crc kubenswrapper[4578]: E1003 13:10:37.316555 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="62242ee6-b3ff-4771-a989-4d7bd1d243d3" Oct 03 13:10:37 crc kubenswrapper[4578]: E1003 13:10:37.317995 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="79fcccfa-6df4-41f2-965b-357e5a7984ab" Oct 03 13:10:37 crc kubenswrapper[4578]: I1003 13:10:37.459771 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9jl4l"] Oct 03 13:10:37 crc kubenswrapper[4578]: I1003 13:10:37.613183 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 03 13:10:37 crc kubenswrapper[4578]: W1003 13:10:37.948344 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf1fc342_37e6_4757_b1cf_81a10245997b.slice/crio-1e8d66e8075320c8679c68dca0b0320ad1032861492012704a4fd09b37ccf17c WatchSource:0}: Error finding container 1e8d66e8075320c8679c68dca0b0320ad1032861492012704a4fd09b37ccf17c: Status 404 returned error can't find the container with id 1e8d66e8075320c8679c68dca0b0320ad1032861492012704a4fd09b37ccf17c Oct 03 13:10:37 crc kubenswrapper[4578]: E1003 13:10:37.991160 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Oct 03 13:10:37 crc kubenswrapper[4578]: E1003 13:10:37.991325 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n5ddh58fh67h68h585hcdhd7h564hb9h5c6h9ch5fh77h594h678h64h68bh568h59ch64dh84hf5h84h579h5bfh85h549h59fh687h7ch565h588q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v8v5q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(409ed574-cc6d-49e4-b1b1-4ec50dfe171e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:10:37 crc kubenswrapper[4578]: E1003 13:10:37.992459 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="409ed574-cc6d-49e4-b1b1-4ec50dfe171e" Oct 03 13:10:38 crc kubenswrapper[4578]: I1003 13:10:38.323409 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"df1fc342-37e6-4757-b1cf-81a10245997b","Type":"ContainerStarted","Data":"1e8d66e8075320c8679c68dca0b0320ad1032861492012704a4fd09b37ccf17c"} Oct 03 13:10:38 crc kubenswrapper[4578]: E1003 13:10:38.325194 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="409ed574-cc6d-49e4-b1b1-4ec50dfe171e" Oct 03 13:10:42 crc kubenswrapper[4578]: I1003 13:10:42.353101 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9jl4l" event={"ID":"daabdc14-9eaa-478d-8e0f-3ab92c0568ce","Type":"ContainerStarted","Data":"0253bd25fd0d3185a8c70909ab7c398dc6b512414d9ab4eb862a9bb407fb4e27"} Oct 03 13:10:43 crc kubenswrapper[4578]: E1003 13:10:43.109105 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 03 13:10:43 crc kubenswrapper[4578]: E1003 13:10:43.109557 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jkqt6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-7xdrm_openstack(2a430efe-d234-4742-af6c-40a34b16daf6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:10:43 crc kubenswrapper[4578]: E1003 13:10:43.110816 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" podUID="2a430efe-d234-4742-af6c-40a34b16daf6" Oct 03 13:10:43 crc kubenswrapper[4578]: E1003 13:10:43.125696 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 03 13:10:43 crc kubenswrapper[4578]: E1003 13:10:43.126010 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wrmvg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-5lpwj_openstack(18730769-2ef4-482b-8020-7be7c37f5668): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:10:43 crc kubenswrapper[4578]: E1003 13:10:43.127217 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-5lpwj" podUID="18730769-2ef4-482b-8020-7be7c37f5668" Oct 03 13:10:43 crc kubenswrapper[4578]: E1003 13:10:43.161159 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 03 13:10:43 crc kubenswrapper[4578]: E1003 13:10:43.161502 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kq8jc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-z97ct_openstack(dc1c7337-6ef2-409f-b307-6f1cf60ee599): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:10:43 crc kubenswrapper[4578]: E1003 13:10:43.164872 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" podUID="dc1c7337-6ef2-409f-b307-6f1cf60ee599" Oct 03 13:10:43 crc kubenswrapper[4578]: E1003 13:10:43.168153 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 03 13:10:43 crc kubenswrapper[4578]: E1003 13:10:43.168305 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tz896,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-745tx_openstack(d175b8ea-f1f6-4f08-a577-e596d7024d20): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:10:43 crc kubenswrapper[4578]: E1003 13:10:43.169831 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-745tx" podUID="d175b8ea-f1f6-4f08-a577-e596d7024d20" Oct 03 13:10:43 crc kubenswrapper[4578]: E1003 13:10:43.364730 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" podUID="2a430efe-d234-4742-af6c-40a34b16daf6" Oct 03 13:10:43 crc kubenswrapper[4578]: E1003 13:10:43.365056 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" podUID="dc1c7337-6ef2-409f-b307-6f1cf60ee599" Oct 03 13:10:43 crc kubenswrapper[4578]: I1003 13:10:43.667944 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 03 13:10:43 crc kubenswrapper[4578]: E1003 13:10:43.945806 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Oct 03 13:10:43 crc kubenswrapper[4578]: E1003 13:10:43.945891 4578 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Oct 03 13:10:43 crc kubenswrapper[4578]: E1003 13:10:43.946068 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f8hdw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(58664f3d-6768-4b3b-a96a-6c22c2c7a8c4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 13:10:43 crc kubenswrapper[4578]: E1003 13:10:43.947264 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="58664f3d-6768-4b3b-a96a-6c22c2c7a8c4" Oct 03 13:10:43 crc kubenswrapper[4578]: W1003 13:10:43.949152 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefb87daa_8d73_444a_94ed_9e44006bec59.slice/crio-0e5229b664a3c52e388f6d3669a313b46696e6cf4f40205f416980976c470b56 WatchSource:0}: Error finding container 0e5229b664a3c52e388f6d3669a313b46696e6cf4f40205f416980976c470b56: Status 404 returned error can't find the container with id 0e5229b664a3c52e388f6d3669a313b46696e6cf4f40205f416980976c470b56 Oct 03 13:10:44 crc kubenswrapper[4578]: I1003 13:10:44.000470 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-745tx" Oct 03 13:10:44 crc kubenswrapper[4578]: I1003 13:10:44.064380 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d175b8ea-f1f6-4f08-a577-e596d7024d20-config\") pod \"d175b8ea-f1f6-4f08-a577-e596d7024d20\" (UID: \"d175b8ea-f1f6-4f08-a577-e596d7024d20\") " Oct 03 13:10:44 crc kubenswrapper[4578]: I1003 13:10:44.064491 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tz896\" (UniqueName: \"kubernetes.io/projected/d175b8ea-f1f6-4f08-a577-e596d7024d20-kube-api-access-tz896\") pod \"d175b8ea-f1f6-4f08-a577-e596d7024d20\" (UID: \"d175b8ea-f1f6-4f08-a577-e596d7024d20\") " Oct 03 13:10:44 crc kubenswrapper[4578]: I1003 13:10:44.065075 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d175b8ea-f1f6-4f08-a577-e596d7024d20-config" (OuterVolumeSpecName: "config") pod "d175b8ea-f1f6-4f08-a577-e596d7024d20" (UID: "d175b8ea-f1f6-4f08-a577-e596d7024d20"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:10:44 crc kubenswrapper[4578]: I1003 13:10:44.065272 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d175b8ea-f1f6-4f08-a577-e596d7024d20-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:10:44 crc kubenswrapper[4578]: I1003 13:10:44.071557 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d175b8ea-f1f6-4f08-a577-e596d7024d20-kube-api-access-tz896" (OuterVolumeSpecName: "kube-api-access-tz896") pod "d175b8ea-f1f6-4f08-a577-e596d7024d20" (UID: "d175b8ea-f1f6-4f08-a577-e596d7024d20"). InnerVolumeSpecName "kube-api-access-tz896". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:10:44 crc kubenswrapper[4578]: I1003 13:10:44.166757 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tz896\" (UniqueName: \"kubernetes.io/projected/d175b8ea-f1f6-4f08-a577-e596d7024d20-kube-api-access-tz896\") on node \"crc\" DevicePath \"\"" Oct 03 13:10:44 crc kubenswrapper[4578]: I1003 13:10:44.367474 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-745tx" event={"ID":"d175b8ea-f1f6-4f08-a577-e596d7024d20","Type":"ContainerDied","Data":"5403336d78f97b6de5b7d772387de3b97071a91afc48b9974f8069023437e6d8"} Oct 03 13:10:44 crc kubenswrapper[4578]: I1003 13:10:44.367552 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-745tx" Oct 03 13:10:44 crc kubenswrapper[4578]: I1003 13:10:44.372939 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"efb87daa-8d73-444a-94ed-9e44006bec59","Type":"ContainerStarted","Data":"0e5229b664a3c52e388f6d3669a313b46696e6cf4f40205f416980976c470b56"} Oct 03 13:10:44 crc kubenswrapper[4578]: E1003 13:10:44.375409 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="58664f3d-6768-4b3b-a96a-6c22c2c7a8c4" Oct 03 13:10:44 crc kubenswrapper[4578]: I1003 13:10:44.440612 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-745tx"] Oct 03 13:10:44 crc kubenswrapper[4578]: I1003 13:10:44.446850 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-745tx"] Oct 03 13:10:44 crc kubenswrapper[4578]: I1003 13:10:44.921864 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d175b8ea-f1f6-4f08-a577-e596d7024d20" path="/var/lib/kubelet/pods/d175b8ea-f1f6-4f08-a577-e596d7024d20/volumes" Oct 03 13:10:48 crc kubenswrapper[4578]: I1003 13:10:48.409031 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-5lpwj" event={"ID":"18730769-2ef4-482b-8020-7be7c37f5668","Type":"ContainerDied","Data":"a8f385ac2a77222a09bc46f90230ceebcc75e202daca65a8f77e7396d90b4b77"} Oct 03 13:10:48 crc kubenswrapper[4578]: I1003 13:10:48.409681 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8f385ac2a77222a09bc46f90230ceebcc75e202daca65a8f77e7396d90b4b77" Oct 03 13:10:48 crc kubenswrapper[4578]: I1003 13:10:48.411712 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-5lpwj" Oct 03 13:10:48 crc kubenswrapper[4578]: I1003 13:10:48.534545 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrmvg\" (UniqueName: \"kubernetes.io/projected/18730769-2ef4-482b-8020-7be7c37f5668-kube-api-access-wrmvg\") pod \"18730769-2ef4-482b-8020-7be7c37f5668\" (UID: \"18730769-2ef4-482b-8020-7be7c37f5668\") " Oct 03 13:10:48 crc kubenswrapper[4578]: I1003 13:10:48.534708 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18730769-2ef4-482b-8020-7be7c37f5668-dns-svc\") pod \"18730769-2ef4-482b-8020-7be7c37f5668\" (UID: \"18730769-2ef4-482b-8020-7be7c37f5668\") " Oct 03 13:10:48 crc kubenswrapper[4578]: I1003 13:10:48.534824 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18730769-2ef4-482b-8020-7be7c37f5668-config\") pod \"18730769-2ef4-482b-8020-7be7c37f5668\" (UID: \"18730769-2ef4-482b-8020-7be7c37f5668\") " Oct 03 13:10:48 crc kubenswrapper[4578]: I1003 13:10:48.535218 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18730769-2ef4-482b-8020-7be7c37f5668-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "18730769-2ef4-482b-8020-7be7c37f5668" (UID: "18730769-2ef4-482b-8020-7be7c37f5668"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:10:48 crc kubenswrapper[4578]: I1003 13:10:48.535545 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18730769-2ef4-482b-8020-7be7c37f5668-config" (OuterVolumeSpecName: "config") pod "18730769-2ef4-482b-8020-7be7c37f5668" (UID: "18730769-2ef4-482b-8020-7be7c37f5668"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:10:48 crc kubenswrapper[4578]: I1003 13:10:48.540058 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18730769-2ef4-482b-8020-7be7c37f5668-kube-api-access-wrmvg" (OuterVolumeSpecName: "kube-api-access-wrmvg") pod "18730769-2ef4-482b-8020-7be7c37f5668" (UID: "18730769-2ef4-482b-8020-7be7c37f5668"). InnerVolumeSpecName "kube-api-access-wrmvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:10:48 crc kubenswrapper[4578]: I1003 13:10:48.637033 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18730769-2ef4-482b-8020-7be7c37f5668-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:10:48 crc kubenswrapper[4578]: I1003 13:10:48.637411 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrmvg\" (UniqueName: \"kubernetes.io/projected/18730769-2ef4-482b-8020-7be7c37f5668-kube-api-access-wrmvg\") on node \"crc\" DevicePath \"\"" Oct 03 13:10:48 crc kubenswrapper[4578]: I1003 13:10:48.637437 4578 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18730769-2ef4-482b-8020-7be7c37f5668-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:10:49 crc kubenswrapper[4578]: I1003 13:10:49.415852 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-5lpwj" Oct 03 13:10:49 crc kubenswrapper[4578]: I1003 13:10:49.463318 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-5lpwj"] Oct 03 13:10:49 crc kubenswrapper[4578]: I1003 13:10:49.470329 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-5lpwj"] Oct 03 13:10:50 crc kubenswrapper[4578]: I1003 13:10:50.924204 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18730769-2ef4-482b-8020-7be7c37f5668" path="/var/lib/kubelet/pods/18730769-2ef4-482b-8020-7be7c37f5668/volumes" Oct 03 13:11:00 crc kubenswrapper[4578]: E1003 13:11:00.014097 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified" Oct 03 13:11:00 crc kubenswrapper[4578]: E1003 13:11:00.014772 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovsdbserver-nb,Image:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,Command:[/usr/bin/dumb-init],Args:[/usr/local/bin/container-scripts/setup.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n85h5cch55h565h657hddhf5h56bh79h58dh59ch89hb4h57bh544h8dh544h665hcch5fdh657h65dh65hcdh9h6ch584h674h677h66h5d7h676q,ValueFrom:nil,},EnvVar{Name:OVN_LOGDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovndbcluster-nb-etc-ovn,ReadOnly:false,MountPath:/etc/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9ff8w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/cleanup.sh],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/pidof ovsdb-server],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:20,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(df1fc342-37e6-4757-b1cf-81a10245997b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:11:00 crc kubenswrapper[4578]: I1003 13:11:00.494091 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9jl4l" event={"ID":"daabdc14-9eaa-478d-8e0f-3ab92c0568ce","Type":"ContainerStarted","Data":"a584cdfe7b28b1a4dad0bf73b732537ac80070625176bf64b41a9b551fefaf31"} Oct 03 13:11:00 crc kubenswrapper[4578]: I1003 13:11:00.494539 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-9jl4l" Oct 03 13:11:00 crc kubenswrapper[4578]: I1003 13:11:00.498436 4578 generic.go:334] "Generic (PLEG): container finished" podID="14bc2fd3-f370-48d0-8711-168b0258f21f" containerID="95e5ecd58132a38f8430b85d2a54e1dd82bb7fdb8b8cfa82f8e10a97fa95cf63" exitCode=0 Oct 03 13:11:00 crc kubenswrapper[4578]: I1003 13:11:00.498503 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2sqbb" event={"ID":"14bc2fd3-f370-48d0-8711-168b0258f21f","Type":"ContainerDied","Data":"95e5ecd58132a38f8430b85d2a54e1dd82bb7fdb8b8cfa82f8e10a97fa95cf63"} Oct 03 13:11:00 crc kubenswrapper[4578]: I1003 13:11:00.508269 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"409ed574-cc6d-49e4-b1b1-4ec50dfe171e","Type":"ContainerStarted","Data":"039703a674946782f9e022723ad1d54adf324cd98409e9f075c6603d5a8cf87e"} Oct 03 13:11:00 crc kubenswrapper[4578]: I1003 13:11:00.508492 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 03 13:11:00 crc kubenswrapper[4578]: I1003 13:11:00.513688 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"62242ee6-b3ff-4771-a989-4d7bd1d243d3","Type":"ContainerStarted","Data":"ff2bcc09bcceb4808fcee73bfb7a76e56d50b8ccf6da770279b1edd80b9a9872"} Oct 03 13:11:00 crc kubenswrapper[4578]: I1003 13:11:00.517947 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"32caee81-c7c2-45ee-bb87-2e6003d6115d","Type":"ContainerStarted","Data":"be4cd6452514e6cb3813330e57b4c11c6b476827f9dabb9ed337df6f572b49ee"} Oct 03 13:11:00 crc kubenswrapper[4578]: I1003 13:11:00.528747 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-9jl4l" podStartSLOduration=25.525029246 podStartE2EDuration="42.528727886s" podCreationTimestamp="2025-10-03 13:10:18 +0000 UTC" firstStartedPulling="2025-10-03 13:10:42.341173177 +0000 UTC m=+1178.139645361" lastFinishedPulling="2025-10-03 13:10:59.344871827 +0000 UTC m=+1195.143344001" observedRunningTime="2025-10-03 13:11:00.518504214 +0000 UTC m=+1196.316976408" watchObservedRunningTime="2025-10-03 13:11:00.528727886 +0000 UTC m=+1196.327200100" Oct 03 13:11:00 crc kubenswrapper[4578]: I1003 13:11:00.604066 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=3.392400084 podStartE2EDuration="47.604043722s" podCreationTimestamp="2025-10-03 13:10:13 +0000 UTC" firstStartedPulling="2025-10-03 13:10:15.155966206 +0000 UTC m=+1150.954438390" lastFinishedPulling="2025-10-03 13:10:59.367609844 +0000 UTC m=+1195.166082028" observedRunningTime="2025-10-03 13:11:00.602647317 +0000 UTC m=+1196.401119511" watchObservedRunningTime="2025-10-03 13:11:00.604043722 +0000 UTC m=+1196.402515906" Oct 03 13:11:01 crc kubenswrapper[4578]: I1003 13:11:01.534795 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"79fcccfa-6df4-41f2-965b-357e5a7984ab","Type":"ContainerStarted","Data":"ff3fb681d8b8aa234547e28273c535c33bdeabb40fca2a2d071b337660eef7cd"} Oct 03 13:11:01 crc kubenswrapper[4578]: I1003 13:11:01.542009 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2sqbb" event={"ID":"14bc2fd3-f370-48d0-8711-168b0258f21f","Type":"ContainerStarted","Data":"01f3cbeb88336c7bbb8fdf355ac4f08e9e6ca0cbd46b150fa59754b12f01f186"} Oct 03 13:11:01 crc kubenswrapper[4578]: I1003 13:11:01.542059 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2sqbb" event={"ID":"14bc2fd3-f370-48d0-8711-168b0258f21f","Type":"ContainerStarted","Data":"e418d673de33a6d31570a4aa072bf731311612d1044be1fe396a1d814b8a0f86"} Oct 03 13:11:01 crc kubenswrapper[4578]: I1003 13:11:01.542879 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:11:01 crc kubenswrapper[4578]: I1003 13:11:01.542914 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:11:01 crc kubenswrapper[4578]: I1003 13:11:01.547963 4578 generic.go:334] "Generic (PLEG): container finished" podID="2a430efe-d234-4742-af6c-40a34b16daf6" containerID="d841514305e229a8be7a21be6db60c1fbefdf0a73a21295eb31c4b7636e18cbc" exitCode=0 Oct 03 13:11:01 crc kubenswrapper[4578]: I1003 13:11:01.548043 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" event={"ID":"2a430efe-d234-4742-af6c-40a34b16daf6","Type":"ContainerDied","Data":"d841514305e229a8be7a21be6db60c1fbefdf0a73a21295eb31c4b7636e18cbc"} Oct 03 13:11:01 crc kubenswrapper[4578]: I1003 13:11:01.555923 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c196bfb1-793e-4e04-9602-db06886385ad","Type":"ContainerStarted","Data":"d3c16cfe0b8487b875181d1c37dc0cd7130f888f36465be2bb67dafa2f43ecaa"} Oct 03 13:11:01 crc kubenswrapper[4578]: I1003 13:11:01.575120 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"58664f3d-6768-4b3b-a96a-6c22c2c7a8c4","Type":"ContainerStarted","Data":"48e58c784bfce1386f4309f8d72c8b868d8a7048f75cda2ea774eb2f01a36741"} Oct 03 13:11:01 crc kubenswrapper[4578]: I1003 13:11:01.575390 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 03 13:11:01 crc kubenswrapper[4578]: I1003 13:11:01.589383 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"efb87daa-8d73-444a-94ed-9e44006bec59","Type":"ContainerStarted","Data":"3800a0db874237746aa0527eeae7b42fc465722aaca47405418b9e1e95388cdb"} Oct 03 13:11:01 crc kubenswrapper[4578]: I1003 13:11:01.598395 4578 generic.go:334] "Generic (PLEG): container finished" podID="dc1c7337-6ef2-409f-b307-6f1cf60ee599" containerID="724480aa58b300f7887a99e6f0037558cb44ff644466ed10e8d70b5072d97957" exitCode=0 Oct 03 13:11:01 crc kubenswrapper[4578]: I1003 13:11:01.598579 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" event={"ID":"dc1c7337-6ef2-409f-b307-6f1cf60ee599","Type":"ContainerDied","Data":"724480aa58b300f7887a99e6f0037558cb44ff644466ed10e8d70b5072d97957"} Oct 03 13:11:01 crc kubenswrapper[4578]: I1003 13:11:01.660016 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.8594504130000002 podStartE2EDuration="45.659837551s" podCreationTimestamp="2025-10-03 13:10:16 +0000 UTC" firstStartedPulling="2025-10-03 13:10:17.578491202 +0000 UTC m=+1153.376963386" lastFinishedPulling="2025-10-03 13:11:00.37887834 +0000 UTC m=+1196.177350524" observedRunningTime="2025-10-03 13:11:01.621051407 +0000 UTC m=+1197.419523581" watchObservedRunningTime="2025-10-03 13:11:01.659837551 +0000 UTC m=+1197.458309735" Oct 03 13:11:01 crc kubenswrapper[4578]: I1003 13:11:01.776236 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-2sqbb" podStartSLOduration=22.49633604 podStartE2EDuration="43.776211131s" podCreationTimestamp="2025-10-03 13:10:18 +0000 UTC" firstStartedPulling="2025-10-03 13:10:37.134999084 +0000 UTC m=+1172.933471278" lastFinishedPulling="2025-10-03 13:10:58.414874175 +0000 UTC m=+1194.213346369" observedRunningTime="2025-10-03 13:11:01.745275876 +0000 UTC m=+1197.543748070" watchObservedRunningTime="2025-10-03 13:11:01.776211131 +0000 UTC m=+1197.574683305" Oct 03 13:11:02 crc kubenswrapper[4578]: I1003 13:11:02.610485 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" event={"ID":"2a430efe-d234-4742-af6c-40a34b16daf6","Type":"ContainerStarted","Data":"0f0fa9e6c3606ed02c80167d364abd50b3bfabc42cb3579a90914845619d226b"} Oct 03 13:11:02 crc kubenswrapper[4578]: I1003 13:11:02.611566 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" Oct 03 13:11:02 crc kubenswrapper[4578]: I1003 13:11:02.615410 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" event={"ID":"dc1c7337-6ef2-409f-b307-6f1cf60ee599","Type":"ContainerStarted","Data":"335f718b672fca0c43d2815438c6f138114f8f50b7acdb7b7c957b708f971da9"} Oct 03 13:11:02 crc kubenswrapper[4578]: I1003 13:11:02.632045 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" podStartSLOduration=4.007513607 podStartE2EDuration="54.632030944s" podCreationTimestamp="2025-10-03 13:10:08 +0000 UTC" firstStartedPulling="2025-10-03 13:10:09.671786408 +0000 UTC m=+1145.470258592" lastFinishedPulling="2025-10-03 13:11:00.296303745 +0000 UTC m=+1196.094775929" observedRunningTime="2025-10-03 13:11:02.631420185 +0000 UTC m=+1198.429892369" watchObservedRunningTime="2025-10-03 13:11:02.632030944 +0000 UTC m=+1198.430503128" Oct 03 13:11:04 crc kubenswrapper[4578]: I1003 13:11:04.283567 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" Oct 03 13:11:04 crc kubenswrapper[4578]: I1003 13:11:04.637157 4578 generic.go:334] "Generic (PLEG): container finished" podID="62242ee6-b3ff-4771-a989-4d7bd1d243d3" containerID="ff2bcc09bcceb4808fcee73bfb7a76e56d50b8ccf6da770279b1edd80b9a9872" exitCode=0 Oct 03 13:11:04 crc kubenswrapper[4578]: I1003 13:11:04.637226 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"62242ee6-b3ff-4771-a989-4d7bd1d243d3","Type":"ContainerDied","Data":"ff2bcc09bcceb4808fcee73bfb7a76e56d50b8ccf6da770279b1edd80b9a9872"} Oct 03 13:11:04 crc kubenswrapper[4578]: I1003 13:11:04.645910 4578 generic.go:334] "Generic (PLEG): container finished" podID="32caee81-c7c2-45ee-bb87-2e6003d6115d" containerID="be4cd6452514e6cb3813330e57b4c11c6b476827f9dabb9ed337df6f572b49ee" exitCode=0 Oct 03 13:11:04 crc kubenswrapper[4578]: I1003 13:11:04.645967 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"32caee81-c7c2-45ee-bb87-2e6003d6115d","Type":"ContainerDied","Data":"be4cd6452514e6cb3813330e57b4c11c6b476827f9dabb9ed337df6f572b49ee"} Oct 03 13:11:04 crc kubenswrapper[4578]: I1003 13:11:04.686470 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" podStartSLOduration=6.513790409 podStartE2EDuration="56.68642733s" podCreationTimestamp="2025-10-03 13:10:08 +0000 UTC" firstStartedPulling="2025-10-03 13:10:10.12577883 +0000 UTC m=+1145.924251014" lastFinishedPulling="2025-10-03 13:11:00.298415751 +0000 UTC m=+1196.096887935" observedRunningTime="2025-10-03 13:11:02.653684647 +0000 UTC m=+1198.452156841" watchObservedRunningTime="2025-10-03 13:11:04.68642733 +0000 UTC m=+1200.484899514" Oct 03 13:11:04 crc kubenswrapper[4578]: E1003 13:11:04.797730 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-nb-0" podUID="df1fc342-37e6-4757-b1cf-81a10245997b" Oct 03 13:11:05 crc kubenswrapper[4578]: I1003 13:11:05.655073 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"efb87daa-8d73-444a-94ed-9e44006bec59","Type":"ContainerStarted","Data":"f73288f78e9c83f72324011e7dbbf2b86dfb573f9e291a7d7a8fdcc9058e0edc"} Oct 03 13:11:05 crc kubenswrapper[4578]: I1003 13:11:05.659236 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"62242ee6-b3ff-4771-a989-4d7bd1d243d3","Type":"ContainerStarted","Data":"f8c8963a930ce33d71f251b555fa0902662ac277c1c09d92880a6bf9d06cd808"} Oct 03 13:11:05 crc kubenswrapper[4578]: I1003 13:11:05.662784 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"32caee81-c7c2-45ee-bb87-2e6003d6115d","Type":"ContainerStarted","Data":"a1b61ea22561473f72719701a9b2c9f8b1f265b407b5e3a7813dcb93ae895be4"} Oct 03 13:11:05 crc kubenswrapper[4578]: I1003 13:11:05.666452 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"df1fc342-37e6-4757-b1cf-81a10245997b","Type":"ContainerStarted","Data":"f0b9a119bbf9d455e9752f60ca644dc1f32c1153e663d1babf83670ccabde993"} Oct 03 13:11:05 crc kubenswrapper[4578]: E1003 13:11:05.668071 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="df1fc342-37e6-4757-b1cf-81a10245997b" Oct 03 13:11:05 crc kubenswrapper[4578]: I1003 13:11:05.685242 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=24.103636886 podStartE2EDuration="44.685221602s" podCreationTimestamp="2025-10-03 13:10:21 +0000 UTC" firstStartedPulling="2025-10-03 13:10:43.951094054 +0000 UTC m=+1179.749566238" lastFinishedPulling="2025-10-03 13:11:04.53267877 +0000 UTC m=+1200.331150954" observedRunningTime="2025-10-03 13:11:05.677061705 +0000 UTC m=+1201.475533909" watchObservedRunningTime="2025-10-03 13:11:05.685221602 +0000 UTC m=+1201.483693786" Oct 03 13:11:05 crc kubenswrapper[4578]: I1003 13:11:05.716255 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371983.138536 podStartE2EDuration="53.716238961s" podCreationTimestamp="2025-10-03 13:10:12 +0000 UTC" firstStartedPulling="2025-10-03 13:10:14.379732032 +0000 UTC m=+1150.178204216" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:11:05.716186319 +0000 UTC m=+1201.514658503" watchObservedRunningTime="2025-10-03 13:11:05.716238961 +0000 UTC m=+1201.514711145" Oct 03 13:11:05 crc kubenswrapper[4578]: I1003 13:11:05.734452 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=17.159770696 podStartE2EDuration="53.734432034s" podCreationTimestamp="2025-10-03 13:10:12 +0000 UTC" firstStartedPulling="2025-10-03 13:10:14.669713357 +0000 UTC m=+1150.468185541" lastFinishedPulling="2025-10-03 13:10:51.244374695 +0000 UTC m=+1187.042846879" observedRunningTime="2025-10-03 13:11:05.733129583 +0000 UTC m=+1201.531601787" watchObservedRunningTime="2025-10-03 13:11:05.734432034 +0000 UTC m=+1201.532904218" Oct 03 13:11:06 crc kubenswrapper[4578]: E1003 13:11:06.674604 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovsdbserver-nb\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="df1fc342-37e6-4757-b1cf-81a10245997b" Oct 03 13:11:06 crc kubenswrapper[4578]: I1003 13:11:06.841500 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 03 13:11:08 crc kubenswrapper[4578]: I1003 13:11:08.237534 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 03 13:11:08 crc kubenswrapper[4578]: I1003 13:11:08.237813 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 03 13:11:08 crc kubenswrapper[4578]: I1003 13:11:08.277625 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 03 13:11:08 crc kubenswrapper[4578]: I1003 13:11:08.733930 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 03 13:11:08 crc kubenswrapper[4578]: I1003 13:11:08.884609 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.032349 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.075261 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-z97ct"] Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.075612 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" podUID="dc1c7337-6ef2-409f-b307-6f1cf60ee599" containerName="dnsmasq-dns" containerID="cri-o://335f718b672fca0c43d2815438c6f138114f8f50b7acdb7b7c957b708f971da9" gracePeriod=10 Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.076840 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.179172 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-9qtv4"] Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.181676 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-9qtv4" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.192950 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.219699 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-bl5dj"] Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.221094 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-bl5dj" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.226043 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.259464 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-9qtv4"] Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.284417 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" podUID="dc1c7337-6ef2-409f-b307-6f1cf60ee599" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.96:5353: connect: connection refused" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.291591 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a36a581c-a79c-44db-9cba-b973af01d8fb-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-9qtv4\" (UID: \"a36a581c-a79c-44db-9cba-b973af01d8fb\") " pod="openstack/dnsmasq-dns-7f896c8c65-9qtv4" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.291680 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a36a581c-a79c-44db-9cba-b973af01d8fb-config\") pod \"dnsmasq-dns-7f896c8c65-9qtv4\" (UID: \"a36a581c-a79c-44db-9cba-b973af01d8fb\") " pod="openstack/dnsmasq-dns-7f896c8c65-9qtv4" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.291947 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmrcp\" (UniqueName: \"kubernetes.io/projected/a36a581c-a79c-44db-9cba-b973af01d8fb-kube-api-access-nmrcp\") pod \"dnsmasq-dns-7f896c8c65-9qtv4\" (UID: \"a36a581c-a79c-44db-9cba-b973af01d8fb\") " pod="openstack/dnsmasq-dns-7f896c8c65-9qtv4" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.292072 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a36a581c-a79c-44db-9cba-b973af01d8fb-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-9qtv4\" (UID: \"a36a581c-a79c-44db-9cba-b973af01d8fb\") " pod="openstack/dnsmasq-dns-7f896c8c65-9qtv4" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.364303 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-bl5dj"] Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.395527 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a36a581c-a79c-44db-9cba-b973af01d8fb-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-9qtv4\" (UID: \"a36a581c-a79c-44db-9cba-b973af01d8fb\") " pod="openstack/dnsmasq-dns-7f896c8c65-9qtv4" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.395597 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2ac114a-65da-41e3-9067-f076c7fbb9c6-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-bl5dj\" (UID: \"d2ac114a-65da-41e3-9067-f076c7fbb9c6\") " pod="openstack/ovn-controller-metrics-bl5dj" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.395628 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d2ac114a-65da-41e3-9067-f076c7fbb9c6-ovn-rundir\") pod \"ovn-controller-metrics-bl5dj\" (UID: \"d2ac114a-65da-41e3-9067-f076c7fbb9c6\") " pod="openstack/ovn-controller-metrics-bl5dj" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.395778 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzpxd\" (UniqueName: \"kubernetes.io/projected/d2ac114a-65da-41e3-9067-f076c7fbb9c6-kube-api-access-rzpxd\") pod \"ovn-controller-metrics-bl5dj\" (UID: \"d2ac114a-65da-41e3-9067-f076c7fbb9c6\") " pod="openstack/ovn-controller-metrics-bl5dj" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.395807 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2ac114a-65da-41e3-9067-f076c7fbb9c6-combined-ca-bundle\") pod \"ovn-controller-metrics-bl5dj\" (UID: \"d2ac114a-65da-41e3-9067-f076c7fbb9c6\") " pod="openstack/ovn-controller-metrics-bl5dj" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.395835 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a36a581c-a79c-44db-9cba-b973af01d8fb-config\") pod \"dnsmasq-dns-7f896c8c65-9qtv4\" (UID: \"a36a581c-a79c-44db-9cba-b973af01d8fb\") " pod="openstack/dnsmasq-dns-7f896c8c65-9qtv4" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.395887 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2ac114a-65da-41e3-9067-f076c7fbb9c6-config\") pod \"ovn-controller-metrics-bl5dj\" (UID: \"d2ac114a-65da-41e3-9067-f076c7fbb9c6\") " pod="openstack/ovn-controller-metrics-bl5dj" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.395924 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmrcp\" (UniqueName: \"kubernetes.io/projected/a36a581c-a79c-44db-9cba-b973af01d8fb-kube-api-access-nmrcp\") pod \"dnsmasq-dns-7f896c8c65-9qtv4\" (UID: \"a36a581c-a79c-44db-9cba-b973af01d8fb\") " pod="openstack/dnsmasq-dns-7f896c8c65-9qtv4" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.395961 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a36a581c-a79c-44db-9cba-b973af01d8fb-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-9qtv4\" (UID: \"a36a581c-a79c-44db-9cba-b973af01d8fb\") " pod="openstack/dnsmasq-dns-7f896c8c65-9qtv4" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.395988 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d2ac114a-65da-41e3-9067-f076c7fbb9c6-ovs-rundir\") pod \"ovn-controller-metrics-bl5dj\" (UID: \"d2ac114a-65da-41e3-9067-f076c7fbb9c6\") " pod="openstack/ovn-controller-metrics-bl5dj" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.397619 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a36a581c-a79c-44db-9cba-b973af01d8fb-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-9qtv4\" (UID: \"a36a581c-a79c-44db-9cba-b973af01d8fb\") " pod="openstack/dnsmasq-dns-7f896c8c65-9qtv4" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.398467 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a36a581c-a79c-44db-9cba-b973af01d8fb-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-9qtv4\" (UID: \"a36a581c-a79c-44db-9cba-b973af01d8fb\") " pod="openstack/dnsmasq-dns-7f896c8c65-9qtv4" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.398665 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a36a581c-a79c-44db-9cba-b973af01d8fb-config\") pod \"dnsmasq-dns-7f896c8c65-9qtv4\" (UID: \"a36a581c-a79c-44db-9cba-b973af01d8fb\") " pod="openstack/dnsmasq-dns-7f896c8c65-9qtv4" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.455973 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmrcp\" (UniqueName: \"kubernetes.io/projected/a36a581c-a79c-44db-9cba-b973af01d8fb-kube-api-access-nmrcp\") pod \"dnsmasq-dns-7f896c8c65-9qtv4\" (UID: \"a36a581c-a79c-44db-9cba-b973af01d8fb\") " pod="openstack/dnsmasq-dns-7f896c8c65-9qtv4" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.501067 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2ac114a-65da-41e3-9067-f076c7fbb9c6-config\") pod \"ovn-controller-metrics-bl5dj\" (UID: \"d2ac114a-65da-41e3-9067-f076c7fbb9c6\") " pod="openstack/ovn-controller-metrics-bl5dj" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.501178 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d2ac114a-65da-41e3-9067-f076c7fbb9c6-ovs-rundir\") pod \"ovn-controller-metrics-bl5dj\" (UID: \"d2ac114a-65da-41e3-9067-f076c7fbb9c6\") " pod="openstack/ovn-controller-metrics-bl5dj" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.501223 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2ac114a-65da-41e3-9067-f076c7fbb9c6-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-bl5dj\" (UID: \"d2ac114a-65da-41e3-9067-f076c7fbb9c6\") " pod="openstack/ovn-controller-metrics-bl5dj" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.501240 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d2ac114a-65da-41e3-9067-f076c7fbb9c6-ovn-rundir\") pod \"ovn-controller-metrics-bl5dj\" (UID: \"d2ac114a-65da-41e3-9067-f076c7fbb9c6\") " pod="openstack/ovn-controller-metrics-bl5dj" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.501273 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzpxd\" (UniqueName: \"kubernetes.io/projected/d2ac114a-65da-41e3-9067-f076c7fbb9c6-kube-api-access-rzpxd\") pod \"ovn-controller-metrics-bl5dj\" (UID: \"d2ac114a-65da-41e3-9067-f076c7fbb9c6\") " pod="openstack/ovn-controller-metrics-bl5dj" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.501297 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2ac114a-65da-41e3-9067-f076c7fbb9c6-combined-ca-bundle\") pod \"ovn-controller-metrics-bl5dj\" (UID: \"d2ac114a-65da-41e3-9067-f076c7fbb9c6\") " pod="openstack/ovn-controller-metrics-bl5dj" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.502762 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d2ac114a-65da-41e3-9067-f076c7fbb9c6-ovn-rundir\") pod \"ovn-controller-metrics-bl5dj\" (UID: \"d2ac114a-65da-41e3-9067-f076c7fbb9c6\") " pod="openstack/ovn-controller-metrics-bl5dj" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.502827 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d2ac114a-65da-41e3-9067-f076c7fbb9c6-ovs-rundir\") pod \"ovn-controller-metrics-bl5dj\" (UID: \"d2ac114a-65da-41e3-9067-f076c7fbb9c6\") " pod="openstack/ovn-controller-metrics-bl5dj" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.503851 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2ac114a-65da-41e3-9067-f076c7fbb9c6-config\") pod \"ovn-controller-metrics-bl5dj\" (UID: \"d2ac114a-65da-41e3-9067-f076c7fbb9c6\") " pod="openstack/ovn-controller-metrics-bl5dj" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.519312 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2ac114a-65da-41e3-9067-f076c7fbb9c6-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-bl5dj\" (UID: \"d2ac114a-65da-41e3-9067-f076c7fbb9c6\") " pod="openstack/ovn-controller-metrics-bl5dj" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.536327 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzpxd\" (UniqueName: \"kubernetes.io/projected/d2ac114a-65da-41e3-9067-f076c7fbb9c6-kube-api-access-rzpxd\") pod \"ovn-controller-metrics-bl5dj\" (UID: \"d2ac114a-65da-41e3-9067-f076c7fbb9c6\") " pod="openstack/ovn-controller-metrics-bl5dj" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.543321 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2ac114a-65da-41e3-9067-f076c7fbb9c6-combined-ca-bundle\") pod \"ovn-controller-metrics-bl5dj\" (UID: \"d2ac114a-65da-41e3-9067-f076c7fbb9c6\") " pod="openstack/ovn-controller-metrics-bl5dj" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.561378 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-9qtv4" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.562575 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-bl5dj" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.728959 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-9qtv4"] Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.738230 4578 generic.go:334] "Generic (PLEG): container finished" podID="dc1c7337-6ef2-409f-b307-6f1cf60ee599" containerID="335f718b672fca0c43d2815438c6f138114f8f50b7acdb7b7c957b708f971da9" exitCode=0 Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.738966 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" event={"ID":"dc1c7337-6ef2-409f-b307-6f1cf60ee599","Type":"ContainerDied","Data":"335f718b672fca0c43d2815438c6f138114f8f50b7acdb7b7c957b708f971da9"} Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.791909 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-p8vpr"] Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.798449 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.813084 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.820106 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-p8vpr"] Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.913462 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-config\") pod \"dnsmasq-dns-86db49b7ff-p8vpr\" (UID: \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\") " pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.913503 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-p8vpr\" (UID: \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\") " pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.913549 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-959j8\" (UniqueName: \"kubernetes.io/projected/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-kube-api-access-959j8\") pod \"dnsmasq-dns-86db49b7ff-p8vpr\" (UID: \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\") " pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.913600 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-p8vpr\" (UID: \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\") " pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.913617 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-p8vpr\" (UID: \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\") " pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" Oct 03 13:11:09 crc kubenswrapper[4578]: I1003 13:11:09.986797 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.018289 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-config\") pod \"dnsmasq-dns-86db49b7ff-p8vpr\" (UID: \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\") " pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.018380 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-p8vpr\" (UID: \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\") " pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.018492 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-959j8\" (UniqueName: \"kubernetes.io/projected/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-kube-api-access-959j8\") pod \"dnsmasq-dns-86db49b7ff-p8vpr\" (UID: \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\") " pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.023387 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-p8vpr\" (UID: \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\") " pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.024029 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-config\") pod \"dnsmasq-dns-86db49b7ff-p8vpr\" (UID: \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\") " pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.025278 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-p8vpr\" (UID: \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\") " pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.026092 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-p8vpr\" (UID: \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\") " pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.026167 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-p8vpr\" (UID: \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\") " pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.028980 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-p8vpr\" (UID: \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\") " pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.058133 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-959j8\" (UniqueName: \"kubernetes.io/projected/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-kube-api-access-959j8\") pod \"dnsmasq-dns-86db49b7ff-p8vpr\" (UID: \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\") " pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.127332 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc1c7337-6ef2-409f-b307-6f1cf60ee599-config\") pod \"dc1c7337-6ef2-409f-b307-6f1cf60ee599\" (UID: \"dc1c7337-6ef2-409f-b307-6f1cf60ee599\") " Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.127508 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kq8jc\" (UniqueName: \"kubernetes.io/projected/dc1c7337-6ef2-409f-b307-6f1cf60ee599-kube-api-access-kq8jc\") pod \"dc1c7337-6ef2-409f-b307-6f1cf60ee599\" (UID: \"dc1c7337-6ef2-409f-b307-6f1cf60ee599\") " Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.127543 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc1c7337-6ef2-409f-b307-6f1cf60ee599-dns-svc\") pod \"dc1c7337-6ef2-409f-b307-6f1cf60ee599\" (UID: \"dc1c7337-6ef2-409f-b307-6f1cf60ee599\") " Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.134854 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc1c7337-6ef2-409f-b307-6f1cf60ee599-kube-api-access-kq8jc" (OuterVolumeSpecName: "kube-api-access-kq8jc") pod "dc1c7337-6ef2-409f-b307-6f1cf60ee599" (UID: "dc1c7337-6ef2-409f-b307-6f1cf60ee599"). InnerVolumeSpecName "kube-api-access-kq8jc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.164766 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc1c7337-6ef2-409f-b307-6f1cf60ee599-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dc1c7337-6ef2-409f-b307-6f1cf60ee599" (UID: "dc1c7337-6ef2-409f-b307-6f1cf60ee599"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.172098 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.173318 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc1c7337-6ef2-409f-b307-6f1cf60ee599-config" (OuterVolumeSpecName: "config") pod "dc1c7337-6ef2-409f-b307-6f1cf60ee599" (UID: "dc1c7337-6ef2-409f-b307-6f1cf60ee599"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.230770 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kq8jc\" (UniqueName: \"kubernetes.io/projected/dc1c7337-6ef2-409f-b307-6f1cf60ee599-kube-api-access-kq8jc\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.230805 4578 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc1c7337-6ef2-409f-b307-6f1cf60ee599-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.230907 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc1c7337-6ef2-409f-b307-6f1cf60ee599-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.341695 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-9qtv4"] Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.448374 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-bl5dj"] Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.509362 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-p8vpr"] Oct 03 13:11:10 crc kubenswrapper[4578]: W1003 13:11:10.540135 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26c21fb1_0b95_4a13_9e50_4cb9aee0202e.slice/crio-317970f7934419387562d1ad5204716c6c9b249331dd108a71566c0005c25fb7 WatchSource:0}: Error finding container 317970f7934419387562d1ad5204716c6c9b249331dd108a71566c0005c25fb7: Status 404 returned error can't find the container with id 317970f7934419387562d1ad5204716c6c9b249331dd108a71566c0005c25fb7 Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.751954 4578 generic.go:334] "Generic (PLEG): container finished" podID="a36a581c-a79c-44db-9cba-b973af01d8fb" containerID="d1831f09bbba979b0be23dd6777f0fe0dadd7ffe51a5a4562f66085c08a555fa" exitCode=0 Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.752053 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-9qtv4" event={"ID":"a36a581c-a79c-44db-9cba-b973af01d8fb","Type":"ContainerDied","Data":"d1831f09bbba979b0be23dd6777f0fe0dadd7ffe51a5a4562f66085c08a555fa"} Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.752737 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-9qtv4" event={"ID":"a36a581c-a79c-44db-9cba-b973af01d8fb","Type":"ContainerStarted","Data":"011b0bb10cffc032e22a39379300801f4c3fdeb1708f1456a504f8c56acc44b5"} Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.754256 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" event={"ID":"26c21fb1-0b95-4a13-9e50-4cb9aee0202e","Type":"ContainerStarted","Data":"317970f7934419387562d1ad5204716c6c9b249331dd108a71566c0005c25fb7"} Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.761570 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-bl5dj" event={"ID":"d2ac114a-65da-41e3-9067-f076c7fbb9c6","Type":"ContainerStarted","Data":"0dd946ce6e860e959ecce1bb8f0688f9b9d8adef9ab496464ead731360e076bc"} Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.765323 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" event={"ID":"dc1c7337-6ef2-409f-b307-6f1cf60ee599","Type":"ContainerDied","Data":"59c8020257a7ae607289ed85e0b110031b3f1c8e5544ecc7598c3ce749a7b790"} Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.765372 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-z97ct" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.765409 4578 scope.go:117] "RemoveContainer" containerID="335f718b672fca0c43d2815438c6f138114f8f50b7acdb7b7c957b708f971da9" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.811254 4578 scope.go:117] "RemoveContainer" containerID="724480aa58b300f7887a99e6f0037558cb44ff644466ed10e8d70b5072d97957" Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.816194 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-z97ct"] Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.820178 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-z97ct"] Oct 03 13:11:10 crc kubenswrapper[4578]: I1003 13:11:10.934897 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc1c7337-6ef2-409f-b307-6f1cf60ee599" path="/var/lib/kubelet/pods/dc1c7337-6ef2-409f-b307-6f1cf60ee599/volumes" Oct 03 13:11:11 crc kubenswrapper[4578]: I1003 13:11:11.059183 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-9qtv4" Oct 03 13:11:11 crc kubenswrapper[4578]: I1003 13:11:11.148091 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a36a581c-a79c-44db-9cba-b973af01d8fb-ovsdbserver-sb\") pod \"a36a581c-a79c-44db-9cba-b973af01d8fb\" (UID: \"a36a581c-a79c-44db-9cba-b973af01d8fb\") " Oct 03 13:11:11 crc kubenswrapper[4578]: I1003 13:11:11.148495 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmrcp\" (UniqueName: \"kubernetes.io/projected/a36a581c-a79c-44db-9cba-b973af01d8fb-kube-api-access-nmrcp\") pod \"a36a581c-a79c-44db-9cba-b973af01d8fb\" (UID: \"a36a581c-a79c-44db-9cba-b973af01d8fb\") " Oct 03 13:11:11 crc kubenswrapper[4578]: I1003 13:11:11.148531 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a36a581c-a79c-44db-9cba-b973af01d8fb-config\") pod \"a36a581c-a79c-44db-9cba-b973af01d8fb\" (UID: \"a36a581c-a79c-44db-9cba-b973af01d8fb\") " Oct 03 13:11:11 crc kubenswrapper[4578]: I1003 13:11:11.148552 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a36a581c-a79c-44db-9cba-b973af01d8fb-dns-svc\") pod \"a36a581c-a79c-44db-9cba-b973af01d8fb\" (UID: \"a36a581c-a79c-44db-9cba-b973af01d8fb\") " Oct 03 13:11:11 crc kubenswrapper[4578]: I1003 13:11:11.155993 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a36a581c-a79c-44db-9cba-b973af01d8fb-kube-api-access-nmrcp" (OuterVolumeSpecName: "kube-api-access-nmrcp") pod "a36a581c-a79c-44db-9cba-b973af01d8fb" (UID: "a36a581c-a79c-44db-9cba-b973af01d8fb"). InnerVolumeSpecName "kube-api-access-nmrcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:11:11 crc kubenswrapper[4578]: I1003 13:11:11.168666 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a36a581c-a79c-44db-9cba-b973af01d8fb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a36a581c-a79c-44db-9cba-b973af01d8fb" (UID: "a36a581c-a79c-44db-9cba-b973af01d8fb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:11:11 crc kubenswrapper[4578]: I1003 13:11:11.169326 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a36a581c-a79c-44db-9cba-b973af01d8fb-config" (OuterVolumeSpecName: "config") pod "a36a581c-a79c-44db-9cba-b973af01d8fb" (UID: "a36a581c-a79c-44db-9cba-b973af01d8fb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:11:11 crc kubenswrapper[4578]: I1003 13:11:11.180571 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a36a581c-a79c-44db-9cba-b973af01d8fb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a36a581c-a79c-44db-9cba-b973af01d8fb" (UID: "a36a581c-a79c-44db-9cba-b973af01d8fb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:11:11 crc kubenswrapper[4578]: I1003 13:11:11.251060 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a36a581c-a79c-44db-9cba-b973af01d8fb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:11 crc kubenswrapper[4578]: I1003 13:11:11.251101 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmrcp\" (UniqueName: \"kubernetes.io/projected/a36a581c-a79c-44db-9cba-b973af01d8fb-kube-api-access-nmrcp\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:11 crc kubenswrapper[4578]: I1003 13:11:11.251115 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a36a581c-a79c-44db-9cba-b973af01d8fb-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:11 crc kubenswrapper[4578]: I1003 13:11:11.251126 4578 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a36a581c-a79c-44db-9cba-b973af01d8fb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:11 crc kubenswrapper[4578]: I1003 13:11:11.773499 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-bl5dj" event={"ID":"d2ac114a-65da-41e3-9067-f076c7fbb9c6","Type":"ContainerStarted","Data":"7e3edfe6a9ae145d7bde86d9ceb05c035b6a163891369fdb2fbd8671f871e79f"} Oct 03 13:11:11 crc kubenswrapper[4578]: I1003 13:11:11.777289 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-9qtv4" event={"ID":"a36a581c-a79c-44db-9cba-b973af01d8fb","Type":"ContainerDied","Data":"011b0bb10cffc032e22a39379300801f4c3fdeb1708f1456a504f8c56acc44b5"} Oct 03 13:11:11 crc kubenswrapper[4578]: I1003 13:11:11.777337 4578 scope.go:117] "RemoveContainer" containerID="d1831f09bbba979b0be23dd6777f0fe0dadd7ffe51a5a4562f66085c08a555fa" Oct 03 13:11:11 crc kubenswrapper[4578]: I1003 13:11:11.777451 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-9qtv4" Oct 03 13:11:11 crc kubenswrapper[4578]: I1003 13:11:11.781228 4578 generic.go:334] "Generic (PLEG): container finished" podID="26c21fb1-0b95-4a13-9e50-4cb9aee0202e" containerID="e0809687229017a0c85fa6abfe5343074633591a8299139cd96a44aed69ee308" exitCode=0 Oct 03 13:11:11 crc kubenswrapper[4578]: I1003 13:11:11.781324 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" event={"ID":"26c21fb1-0b95-4a13-9e50-4cb9aee0202e","Type":"ContainerDied","Data":"e0809687229017a0c85fa6abfe5343074633591a8299139cd96a44aed69ee308"} Oct 03 13:11:11 crc kubenswrapper[4578]: I1003 13:11:11.799611 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-bl5dj" podStartSLOduration=2.799580939 podStartE2EDuration="2.799580939s" podCreationTimestamp="2025-10-03 13:11:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:11:11.794159709 +0000 UTC m=+1207.592631893" watchObservedRunningTime="2025-10-03 13:11:11.799580939 +0000 UTC m=+1207.598053163" Oct 03 13:11:12 crc kubenswrapper[4578]: I1003 13:11:12.000216 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-9qtv4"] Oct 03 13:11:12 crc kubenswrapper[4578]: I1003 13:11:12.015746 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-9qtv4"] Oct 03 13:11:12 crc kubenswrapper[4578]: I1003 13:11:12.792234 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" event={"ID":"26c21fb1-0b95-4a13-9e50-4cb9aee0202e","Type":"ContainerStarted","Data":"0984de7bd3a8ddee591db0b6a8511bb2294437ebcad030244875009d58db7bd3"} Oct 03 13:11:12 crc kubenswrapper[4578]: I1003 13:11:12.793200 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" Oct 03 13:11:12 crc kubenswrapper[4578]: I1003 13:11:12.815033 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" podStartSLOduration=3.815011056 podStartE2EDuration="3.815011056s" podCreationTimestamp="2025-10-03 13:11:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:11:12.807794608 +0000 UTC m=+1208.606266802" watchObservedRunningTime="2025-10-03 13:11:12.815011056 +0000 UTC m=+1208.613483230" Oct 03 13:11:12 crc kubenswrapper[4578]: I1003 13:11:12.918807 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a36a581c-a79c-44db-9cba-b973af01d8fb" path="/var/lib/kubelet/pods/a36a581c-a79c-44db-9cba-b973af01d8fb/volumes" Oct 03 13:11:13 crc kubenswrapper[4578]: I1003 13:11:13.386977 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 03 13:11:13 crc kubenswrapper[4578]: I1003 13:11:13.387507 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 03 13:11:13 crc kubenswrapper[4578]: I1003 13:11:13.433336 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 03 13:11:13 crc kubenswrapper[4578]: I1003 13:11:13.740704 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 03 13:11:13 crc kubenswrapper[4578]: I1003 13:11:13.740775 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 03 13:11:13 crc kubenswrapper[4578]: I1003 13:11:13.801727 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 03 13:11:13 crc kubenswrapper[4578]: I1003 13:11:13.865103 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 03 13:11:13 crc kubenswrapper[4578]: I1003 13:11:13.867415 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 03 13:11:14 crc kubenswrapper[4578]: I1003 13:11:14.088260 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-6l7bj"] Oct 03 13:11:14 crc kubenswrapper[4578]: E1003 13:11:14.089008 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a36a581c-a79c-44db-9cba-b973af01d8fb" containerName="init" Oct 03 13:11:14 crc kubenswrapper[4578]: I1003 13:11:14.089026 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a36a581c-a79c-44db-9cba-b973af01d8fb" containerName="init" Oct 03 13:11:14 crc kubenswrapper[4578]: E1003 13:11:14.089037 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc1c7337-6ef2-409f-b307-6f1cf60ee599" containerName="init" Oct 03 13:11:14 crc kubenswrapper[4578]: I1003 13:11:14.089073 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc1c7337-6ef2-409f-b307-6f1cf60ee599" containerName="init" Oct 03 13:11:14 crc kubenswrapper[4578]: E1003 13:11:14.089090 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc1c7337-6ef2-409f-b307-6f1cf60ee599" containerName="dnsmasq-dns" Oct 03 13:11:14 crc kubenswrapper[4578]: I1003 13:11:14.089096 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc1c7337-6ef2-409f-b307-6f1cf60ee599" containerName="dnsmasq-dns" Oct 03 13:11:14 crc kubenswrapper[4578]: I1003 13:11:14.089325 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc1c7337-6ef2-409f-b307-6f1cf60ee599" containerName="dnsmasq-dns" Oct 03 13:11:14 crc kubenswrapper[4578]: I1003 13:11:14.089342 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="a36a581c-a79c-44db-9cba-b973af01d8fb" containerName="init" Oct 03 13:11:14 crc kubenswrapper[4578]: I1003 13:11:14.089945 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-6l7bj" Oct 03 13:11:14 crc kubenswrapper[4578]: I1003 13:11:14.095532 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-6l7bj"] Oct 03 13:11:14 crc kubenswrapper[4578]: I1003 13:11:14.202938 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ndcz\" (UniqueName: \"kubernetes.io/projected/c8d4460d-feb5-4d93-95e4-260f5e06ee93-kube-api-access-7ndcz\") pod \"placement-db-create-6l7bj\" (UID: \"c8d4460d-feb5-4d93-95e4-260f5e06ee93\") " pod="openstack/placement-db-create-6l7bj" Oct 03 13:11:14 crc kubenswrapper[4578]: I1003 13:11:14.305003 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ndcz\" (UniqueName: \"kubernetes.io/projected/c8d4460d-feb5-4d93-95e4-260f5e06ee93-kube-api-access-7ndcz\") pod \"placement-db-create-6l7bj\" (UID: \"c8d4460d-feb5-4d93-95e4-260f5e06ee93\") " pod="openstack/placement-db-create-6l7bj" Oct 03 13:11:14 crc kubenswrapper[4578]: I1003 13:11:14.327878 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ndcz\" (UniqueName: \"kubernetes.io/projected/c8d4460d-feb5-4d93-95e4-260f5e06ee93-kube-api-access-7ndcz\") pod \"placement-db-create-6l7bj\" (UID: \"c8d4460d-feb5-4d93-95e4-260f5e06ee93\") " pod="openstack/placement-db-create-6l7bj" Oct 03 13:11:14 crc kubenswrapper[4578]: I1003 13:11:14.409108 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-6l7bj" Oct 03 13:11:14 crc kubenswrapper[4578]: I1003 13:11:14.829522 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-6l7bj"] Oct 03 13:11:15 crc kubenswrapper[4578]: I1003 13:11:15.823893 4578 generic.go:334] "Generic (PLEG): container finished" podID="c8d4460d-feb5-4d93-95e4-260f5e06ee93" containerID="a4c1d9d3955644651e82ef57819ea32c343727a4ac98c390148a5c47d256e96e" exitCode=0 Oct 03 13:11:15 crc kubenswrapper[4578]: I1003 13:11:15.823967 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-6l7bj" event={"ID":"c8d4460d-feb5-4d93-95e4-260f5e06ee93","Type":"ContainerDied","Data":"a4c1d9d3955644651e82ef57819ea32c343727a4ac98c390148a5c47d256e96e"} Oct 03 13:11:15 crc kubenswrapper[4578]: I1003 13:11:15.824379 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-6l7bj" event={"ID":"c8d4460d-feb5-4d93-95e4-260f5e06ee93","Type":"ContainerStarted","Data":"5217779913afbc4fee48351994d71111ecc7388945db50baf1cb378792764a2c"} Oct 03 13:11:16 crc kubenswrapper[4578]: I1003 13:11:16.820680 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-p8vpr"] Oct 03 13:11:16 crc kubenswrapper[4578]: I1003 13:11:16.821418 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" podUID="26c21fb1-0b95-4a13-9e50-4cb9aee0202e" containerName="dnsmasq-dns" containerID="cri-o://0984de7bd3a8ddee591db0b6a8511bb2294437ebcad030244875009d58db7bd3" gracePeriod=10 Oct 03 13:11:16 crc kubenswrapper[4578]: I1003 13:11:16.879083 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-k9bht"] Oct 03 13:11:16 crc kubenswrapper[4578]: I1003 13:11:16.880351 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:11:16 crc kubenswrapper[4578]: I1003 13:11:16.898667 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-k9bht"] Oct 03 13:11:16 crc kubenswrapper[4578]: I1003 13:11:16.948942 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-k9bht\" (UID: \"68090354-cfd7-472a-bca7-a9b06f40957f\") " pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:11:16 crc kubenswrapper[4578]: I1003 13:11:16.948992 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4l8k\" (UniqueName: \"kubernetes.io/projected/68090354-cfd7-472a-bca7-a9b06f40957f-kube-api-access-v4l8k\") pod \"dnsmasq-dns-698758b865-k9bht\" (UID: \"68090354-cfd7-472a-bca7-a9b06f40957f\") " pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:11:16 crc kubenswrapper[4578]: I1003 13:11:16.949019 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-k9bht\" (UID: \"68090354-cfd7-472a-bca7-a9b06f40957f\") " pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:11:16 crc kubenswrapper[4578]: I1003 13:11:16.949088 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-dns-svc\") pod \"dnsmasq-dns-698758b865-k9bht\" (UID: \"68090354-cfd7-472a-bca7-a9b06f40957f\") " pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:11:16 crc kubenswrapper[4578]: I1003 13:11:16.949155 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-config\") pod \"dnsmasq-dns-698758b865-k9bht\" (UID: \"68090354-cfd7-472a-bca7-a9b06f40957f\") " pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.051886 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-config\") pod \"dnsmasq-dns-698758b865-k9bht\" (UID: \"68090354-cfd7-472a-bca7-a9b06f40957f\") " pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.051991 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-k9bht\" (UID: \"68090354-cfd7-472a-bca7-a9b06f40957f\") " pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.052013 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4l8k\" (UniqueName: \"kubernetes.io/projected/68090354-cfd7-472a-bca7-a9b06f40957f-kube-api-access-v4l8k\") pod \"dnsmasq-dns-698758b865-k9bht\" (UID: \"68090354-cfd7-472a-bca7-a9b06f40957f\") " pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.052042 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-k9bht\" (UID: \"68090354-cfd7-472a-bca7-a9b06f40957f\") " pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.052096 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-dns-svc\") pod \"dnsmasq-dns-698758b865-k9bht\" (UID: \"68090354-cfd7-472a-bca7-a9b06f40957f\") " pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.053042 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-dns-svc\") pod \"dnsmasq-dns-698758b865-k9bht\" (UID: \"68090354-cfd7-472a-bca7-a9b06f40957f\") " pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.053795 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-k9bht\" (UID: \"68090354-cfd7-472a-bca7-a9b06f40957f\") " pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.053997 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-k9bht\" (UID: \"68090354-cfd7-472a-bca7-a9b06f40957f\") " pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.055572 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-config\") pod \"dnsmasq-dns-698758b865-k9bht\" (UID: \"68090354-cfd7-472a-bca7-a9b06f40957f\") " pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.087609 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4l8k\" (UniqueName: \"kubernetes.io/projected/68090354-cfd7-472a-bca7-a9b06f40957f-kube-api-access-v4l8k\") pod \"dnsmasq-dns-698758b865-k9bht\" (UID: \"68090354-cfd7-472a-bca7-a9b06f40957f\") " pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.215853 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.387703 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-6l7bj" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.468518 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ndcz\" (UniqueName: \"kubernetes.io/projected/c8d4460d-feb5-4d93-95e4-260f5e06ee93-kube-api-access-7ndcz\") pod \"c8d4460d-feb5-4d93-95e4-260f5e06ee93\" (UID: \"c8d4460d-feb5-4d93-95e4-260f5e06ee93\") " Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.485918 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8d4460d-feb5-4d93-95e4-260f5e06ee93-kube-api-access-7ndcz" (OuterVolumeSpecName: "kube-api-access-7ndcz") pod "c8d4460d-feb5-4d93-95e4-260f5e06ee93" (UID: "c8d4460d-feb5-4d93-95e4-260f5e06ee93"). InnerVolumeSpecName "kube-api-access-7ndcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.529948 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.572172 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-dns-svc\") pod \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\" (UID: \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\") " Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.572252 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-config\") pod \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\" (UID: \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\") " Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.572279 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-959j8\" (UniqueName: \"kubernetes.io/projected/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-kube-api-access-959j8\") pod \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\" (UID: \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\") " Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.572382 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-ovsdbserver-sb\") pod \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\" (UID: \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\") " Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.572493 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-ovsdbserver-nb\") pod \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\" (UID: \"26c21fb1-0b95-4a13-9e50-4cb9aee0202e\") " Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.572879 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ndcz\" (UniqueName: \"kubernetes.io/projected/c8d4460d-feb5-4d93-95e4-260f5e06ee93-kube-api-access-7ndcz\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.578564 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-kube-api-access-959j8" (OuterVolumeSpecName: "kube-api-access-959j8") pod "26c21fb1-0b95-4a13-9e50-4cb9aee0202e" (UID: "26c21fb1-0b95-4a13-9e50-4cb9aee0202e"). InnerVolumeSpecName "kube-api-access-959j8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.617282 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "26c21fb1-0b95-4a13-9e50-4cb9aee0202e" (UID: "26c21fb1-0b95-4a13-9e50-4cb9aee0202e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.618085 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "26c21fb1-0b95-4a13-9e50-4cb9aee0202e" (UID: "26c21fb1-0b95-4a13-9e50-4cb9aee0202e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.629987 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "26c21fb1-0b95-4a13-9e50-4cb9aee0202e" (UID: "26c21fb1-0b95-4a13-9e50-4cb9aee0202e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.632088 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-config" (OuterVolumeSpecName: "config") pod "26c21fb1-0b95-4a13-9e50-4cb9aee0202e" (UID: "26c21fb1-0b95-4a13-9e50-4cb9aee0202e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.674820 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.674856 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-959j8\" (UniqueName: \"kubernetes.io/projected/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-kube-api-access-959j8\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.674868 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.674876 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.674884 4578 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26c21fb1-0b95-4a13-9e50-4cb9aee0202e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.844969 4578 generic.go:334] "Generic (PLEG): container finished" podID="26c21fb1-0b95-4a13-9e50-4cb9aee0202e" containerID="0984de7bd3a8ddee591db0b6a8511bb2294437ebcad030244875009d58db7bd3" exitCode=0 Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.845034 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" event={"ID":"26c21fb1-0b95-4a13-9e50-4cb9aee0202e","Type":"ContainerDied","Data":"0984de7bd3a8ddee591db0b6a8511bb2294437ebcad030244875009d58db7bd3"} Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.845060 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" event={"ID":"26c21fb1-0b95-4a13-9e50-4cb9aee0202e","Type":"ContainerDied","Data":"317970f7934419387562d1ad5204716c6c9b249331dd108a71566c0005c25fb7"} Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.845077 4578 scope.go:117] "RemoveContainer" containerID="0984de7bd3a8ddee591db0b6a8511bb2294437ebcad030244875009d58db7bd3" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.845034 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-p8vpr" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.847728 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-6l7bj" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.847750 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-6l7bj" event={"ID":"c8d4460d-feb5-4d93-95e4-260f5e06ee93","Type":"ContainerDied","Data":"5217779913afbc4fee48351994d71111ecc7388945db50baf1cb378792764a2c"} Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.847801 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5217779913afbc4fee48351994d71111ecc7388945db50baf1cb378792764a2c" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.863500 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-k9bht"] Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.869077 4578 scope.go:117] "RemoveContainer" containerID="e0809687229017a0c85fa6abfe5343074633591a8299139cd96a44aed69ee308" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.884022 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-p8vpr"] Oct 03 13:11:17 crc kubenswrapper[4578]: W1003 13:11:17.889748 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68090354_cfd7_472a_bca7_a9b06f40957f.slice/crio-48f8191f2b4fcb115efbc5d51eaa83d11d80093fbc65c5e774f8a97fddce4531 WatchSource:0}: Error finding container 48f8191f2b4fcb115efbc5d51eaa83d11d80093fbc65c5e774f8a97fddce4531: Status 404 returned error can't find the container with id 48f8191f2b4fcb115efbc5d51eaa83d11d80093fbc65c5e774f8a97fddce4531 Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.890992 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-p8vpr"] Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.896892 4578 scope.go:117] "RemoveContainer" containerID="0984de7bd3a8ddee591db0b6a8511bb2294437ebcad030244875009d58db7bd3" Oct 03 13:11:17 crc kubenswrapper[4578]: E1003 13:11:17.897432 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0984de7bd3a8ddee591db0b6a8511bb2294437ebcad030244875009d58db7bd3\": container with ID starting with 0984de7bd3a8ddee591db0b6a8511bb2294437ebcad030244875009d58db7bd3 not found: ID does not exist" containerID="0984de7bd3a8ddee591db0b6a8511bb2294437ebcad030244875009d58db7bd3" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.897460 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0984de7bd3a8ddee591db0b6a8511bb2294437ebcad030244875009d58db7bd3"} err="failed to get container status \"0984de7bd3a8ddee591db0b6a8511bb2294437ebcad030244875009d58db7bd3\": rpc error: code = NotFound desc = could not find container \"0984de7bd3a8ddee591db0b6a8511bb2294437ebcad030244875009d58db7bd3\": container with ID starting with 0984de7bd3a8ddee591db0b6a8511bb2294437ebcad030244875009d58db7bd3 not found: ID does not exist" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.897481 4578 scope.go:117] "RemoveContainer" containerID="e0809687229017a0c85fa6abfe5343074633591a8299139cd96a44aed69ee308" Oct 03 13:11:17 crc kubenswrapper[4578]: E1003 13:11:17.897757 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0809687229017a0c85fa6abfe5343074633591a8299139cd96a44aed69ee308\": container with ID starting with e0809687229017a0c85fa6abfe5343074633591a8299139cd96a44aed69ee308 not found: ID does not exist" containerID="e0809687229017a0c85fa6abfe5343074633591a8299139cd96a44aed69ee308" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.897778 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0809687229017a0c85fa6abfe5343074633591a8299139cd96a44aed69ee308"} err="failed to get container status \"e0809687229017a0c85fa6abfe5343074633591a8299139cd96a44aed69ee308\": rpc error: code = NotFound desc = could not find container \"e0809687229017a0c85fa6abfe5343074633591a8299139cd96a44aed69ee308\": container with ID starting with e0809687229017a0c85fa6abfe5343074633591a8299139cd96a44aed69ee308 not found: ID does not exist" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.988739 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 03 13:11:17 crc kubenswrapper[4578]: E1003 13:11:17.989064 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8d4460d-feb5-4d93-95e4-260f5e06ee93" containerName="mariadb-database-create" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.989076 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8d4460d-feb5-4d93-95e4-260f5e06ee93" containerName="mariadb-database-create" Oct 03 13:11:17 crc kubenswrapper[4578]: E1003 13:11:17.989097 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26c21fb1-0b95-4a13-9e50-4cb9aee0202e" containerName="dnsmasq-dns" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.989103 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="26c21fb1-0b95-4a13-9e50-4cb9aee0202e" containerName="dnsmasq-dns" Oct 03 13:11:17 crc kubenswrapper[4578]: E1003 13:11:17.989115 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26c21fb1-0b95-4a13-9e50-4cb9aee0202e" containerName="init" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.989121 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="26c21fb1-0b95-4a13-9e50-4cb9aee0202e" containerName="init" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.989283 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="26c21fb1-0b95-4a13-9e50-4cb9aee0202e" containerName="dnsmasq-dns" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.989295 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8d4460d-feb5-4d93-95e4-260f5e06ee93" containerName="mariadb-database-create" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.993806 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 03 13:11:17 crc kubenswrapper[4578]: I1003 13:11:17.996255 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.000140 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-96mxh" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.000467 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.002032 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.014364 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.092420 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm87h\" (UniqueName: \"kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-kube-api-access-cm87h\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") " pod="openstack/swift-storage-0" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.092495 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-etc-swift\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") " pod="openstack/swift-storage-0" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.092592 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/aa1d7b22-8dac-42de-8241-b633bbe90691-lock\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") " pod="openstack/swift-storage-0" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.092751 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/aa1d7b22-8dac-42de-8241-b633bbe90691-cache\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") " pod="openstack/swift-storage-0" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.092793 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") " pod="openstack/swift-storage-0" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.193807 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm87h\" (UniqueName: \"kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-kube-api-access-cm87h\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") " pod="openstack/swift-storage-0" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.193865 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-etc-swift\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") " pod="openstack/swift-storage-0" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.193945 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/aa1d7b22-8dac-42de-8241-b633bbe90691-lock\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") " pod="openstack/swift-storage-0" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.194003 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/aa1d7b22-8dac-42de-8241-b633bbe90691-cache\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") " pod="openstack/swift-storage-0" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.194041 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") " pod="openstack/swift-storage-0" Oct 03 13:11:18 crc kubenswrapper[4578]: E1003 13:11:18.194346 4578 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 03 13:11:18 crc kubenswrapper[4578]: E1003 13:11:18.194726 4578 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 03 13:11:18 crc kubenswrapper[4578]: E1003 13:11:18.194862 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-etc-swift podName:aa1d7b22-8dac-42de-8241-b633bbe90691 nodeName:}" failed. No retries permitted until 2025-10-03 13:11:18.694837616 +0000 UTC m=+1214.493309870 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-etc-swift") pod "swift-storage-0" (UID: "aa1d7b22-8dac-42de-8241-b633bbe90691") : configmap "swift-ring-files" not found Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.194469 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/aa1d7b22-8dac-42de-8241-b633bbe90691-cache\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") " pod="openstack/swift-storage-0" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.194373 4578 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/swift-storage-0" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.194439 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/aa1d7b22-8dac-42de-8241-b633bbe90691-lock\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") " pod="openstack/swift-storage-0" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.225947 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") " pod="openstack/swift-storage-0" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.226925 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm87h\" (UniqueName: \"kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-kube-api-access-cm87h\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") " pod="openstack/swift-storage-0" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.232888 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-hjk2v"] Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.235184 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.240406 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.240745 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.240824 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.255068 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-hjk2v"] Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.295097 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2222d5-bfbb-43ca-a40a-d235295823a8-combined-ca-bundle\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.295152 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3f2222d5-bfbb-43ca-a40a-d235295823a8-ring-data-devices\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.295210 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3f2222d5-bfbb-43ca-a40a-d235295823a8-dispersionconf\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.295385 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f2222d5-bfbb-43ca-a40a-d235295823a8-scripts\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.295410 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3f2222d5-bfbb-43ca-a40a-d235295823a8-swiftconf\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.295493 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3f2222d5-bfbb-43ca-a40a-d235295823a8-etc-swift\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.295528 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zmfn\" (UniqueName: \"kubernetes.io/projected/3f2222d5-bfbb-43ca-a40a-d235295823a8-kube-api-access-4zmfn\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.396854 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3f2222d5-bfbb-43ca-a40a-d235295823a8-swiftconf\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.396893 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f2222d5-bfbb-43ca-a40a-d235295823a8-scripts\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.396943 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3f2222d5-bfbb-43ca-a40a-d235295823a8-etc-swift\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.396973 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zmfn\" (UniqueName: \"kubernetes.io/projected/3f2222d5-bfbb-43ca-a40a-d235295823a8-kube-api-access-4zmfn\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.397002 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3f2222d5-bfbb-43ca-a40a-d235295823a8-ring-data-devices\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.397018 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2222d5-bfbb-43ca-a40a-d235295823a8-combined-ca-bundle\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.397033 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3f2222d5-bfbb-43ca-a40a-d235295823a8-dispersionconf\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.397826 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3f2222d5-bfbb-43ca-a40a-d235295823a8-etc-swift\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.400513 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f2222d5-bfbb-43ca-a40a-d235295823a8-scripts\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.401232 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3f2222d5-bfbb-43ca-a40a-d235295823a8-dispersionconf\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.401370 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3f2222d5-bfbb-43ca-a40a-d235295823a8-swiftconf\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.401987 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3f2222d5-bfbb-43ca-a40a-d235295823a8-ring-data-devices\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.419906 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2222d5-bfbb-43ca-a40a-d235295823a8-combined-ca-bundle\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.423122 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zmfn\" (UniqueName: \"kubernetes.io/projected/3f2222d5-bfbb-43ca-a40a-d235295823a8-kube-api-access-4zmfn\") pod \"swift-ring-rebalance-hjk2v\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.601260 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.701439 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-etc-swift\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") " pod="openstack/swift-storage-0" Oct 03 13:11:18 crc kubenswrapper[4578]: E1003 13:11:18.701710 4578 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 03 13:11:18 crc kubenswrapper[4578]: E1003 13:11:18.701930 4578 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 03 13:11:18 crc kubenswrapper[4578]: E1003 13:11:18.701991 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-etc-swift podName:aa1d7b22-8dac-42de-8241-b633bbe90691 nodeName:}" failed. No retries permitted until 2025-10-03 13:11:19.701973972 +0000 UTC m=+1215.500446156 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-etc-swift") pod "swift-storage-0" (UID: "aa1d7b22-8dac-42de-8241-b633bbe90691") : configmap "swift-ring-files" not found Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.862307 4578 generic.go:334] "Generic (PLEG): container finished" podID="68090354-cfd7-472a-bca7-a9b06f40957f" containerID="157d59c22da841c160afb1f51aa48c3d37e8f41c5a8c88abf633a4e7791a7eb1" exitCode=0 Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.862347 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-k9bht" event={"ID":"68090354-cfd7-472a-bca7-a9b06f40957f","Type":"ContainerDied","Data":"157d59c22da841c160afb1f51aa48c3d37e8f41c5a8c88abf633a4e7791a7eb1"} Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.862372 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-k9bht" event={"ID":"68090354-cfd7-472a-bca7-a9b06f40957f","Type":"ContainerStarted","Data":"48f8191f2b4fcb115efbc5d51eaa83d11d80093fbc65c5e774f8a97fddce4531"} Oct 03 13:11:18 crc kubenswrapper[4578]: I1003 13:11:18.918846 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26c21fb1-0b95-4a13-9e50-4cb9aee0202e" path="/var/lib/kubelet/pods/26c21fb1-0b95-4a13-9e50-4cb9aee0202e/volumes" Oct 03 13:11:19 crc kubenswrapper[4578]: I1003 13:11:19.047740 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-hjk2v"] Oct 03 13:11:19 crc kubenswrapper[4578]: I1003 13:11:19.461161 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-xhpcq"] Oct 03 13:11:19 crc kubenswrapper[4578]: I1003 13:11:19.462403 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xhpcq" Oct 03 13:11:19 crc kubenswrapper[4578]: I1003 13:11:19.470776 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-xhpcq"] Oct 03 13:11:19 crc kubenswrapper[4578]: I1003 13:11:19.514714 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4wbn\" (UniqueName: \"kubernetes.io/projected/35460854-45fb-4080-aad3-3638a4283374-kube-api-access-t4wbn\") pod \"glance-db-create-xhpcq\" (UID: \"35460854-45fb-4080-aad3-3638a4283374\") " pod="openstack/glance-db-create-xhpcq" Oct 03 13:11:19 crc kubenswrapper[4578]: I1003 13:11:19.616184 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4wbn\" (UniqueName: \"kubernetes.io/projected/35460854-45fb-4080-aad3-3638a4283374-kube-api-access-t4wbn\") pod \"glance-db-create-xhpcq\" (UID: \"35460854-45fb-4080-aad3-3638a4283374\") " pod="openstack/glance-db-create-xhpcq" Oct 03 13:11:19 crc kubenswrapper[4578]: I1003 13:11:19.635884 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4wbn\" (UniqueName: \"kubernetes.io/projected/35460854-45fb-4080-aad3-3638a4283374-kube-api-access-t4wbn\") pod \"glance-db-create-xhpcq\" (UID: \"35460854-45fb-4080-aad3-3638a4283374\") " pod="openstack/glance-db-create-xhpcq" Oct 03 13:11:19 crc kubenswrapper[4578]: I1003 13:11:19.717544 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-etc-swift\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") " pod="openstack/swift-storage-0" Oct 03 13:11:19 crc kubenswrapper[4578]: E1003 13:11:19.717783 4578 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 03 13:11:19 crc kubenswrapper[4578]: E1003 13:11:19.717816 4578 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 03 13:11:19 crc kubenswrapper[4578]: E1003 13:11:19.717881 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-etc-swift podName:aa1d7b22-8dac-42de-8241-b633bbe90691 nodeName:}" failed. No retries permitted until 2025-10-03 13:11:21.717863452 +0000 UTC m=+1217.516335636 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-etc-swift") pod "swift-storage-0" (UID: "aa1d7b22-8dac-42de-8241-b633bbe90691") : configmap "swift-ring-files" not found Oct 03 13:11:19 crc kubenswrapper[4578]: I1003 13:11:19.798165 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xhpcq" Oct 03 13:11:19 crc kubenswrapper[4578]: I1003 13:11:19.889799 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-k9bht" event={"ID":"68090354-cfd7-472a-bca7-a9b06f40957f","Type":"ContainerStarted","Data":"2ce8a6bad39cb722b50bc5ba9cecd313e4b8e85b77f30dfbe0796b8df15b1ffb"} Oct 03 13:11:19 crc kubenswrapper[4578]: I1003 13:11:19.891089 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:11:19 crc kubenswrapper[4578]: I1003 13:11:19.897946 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hjk2v" event={"ID":"3f2222d5-bfbb-43ca-a40a-d235295823a8","Type":"ContainerStarted","Data":"9d3f4c700f8e6974ef2e67b81b68e3121b8f7048ffc8fa93dfafbe27917ee666"} Oct 03 13:11:19 crc kubenswrapper[4578]: I1003 13:11:19.912586 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-k9bht" podStartSLOduration=3.912564974 podStartE2EDuration="3.912564974s" podCreationTimestamp="2025-10-03 13:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:11:19.907660809 +0000 UTC m=+1215.706133013" watchObservedRunningTime="2025-10-03 13:11:19.912564974 +0000 UTC m=+1215.711037158" Oct 03 13:11:20 crc kubenswrapper[4578]: I1003 13:11:20.221444 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-xhpcq"] Oct 03 13:11:20 crc kubenswrapper[4578]: I1003 13:11:20.907426 4578 generic.go:334] "Generic (PLEG): container finished" podID="35460854-45fb-4080-aad3-3638a4283374" containerID="03a2337d0819cd158802e9abcef1d736a407794a36e8ae2f6f6451d58eae1aeb" exitCode=0 Oct 03 13:11:20 crc kubenswrapper[4578]: I1003 13:11:20.907484 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xhpcq" event={"ID":"35460854-45fb-4080-aad3-3638a4283374","Type":"ContainerDied","Data":"03a2337d0819cd158802e9abcef1d736a407794a36e8ae2f6f6451d58eae1aeb"} Oct 03 13:11:20 crc kubenswrapper[4578]: I1003 13:11:20.907853 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xhpcq" event={"ID":"35460854-45fb-4080-aad3-3638a4283374","Type":"ContainerStarted","Data":"27a3ae700ec4a22e64ac474103569b546f7b11ed7a1b9a02f91ec66140a4da5a"} Oct 03 13:11:21 crc kubenswrapper[4578]: I1003 13:11:21.751226 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-etc-swift\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") " pod="openstack/swift-storage-0" Oct 03 13:11:21 crc kubenswrapper[4578]: E1003 13:11:21.751514 4578 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 03 13:11:21 crc kubenswrapper[4578]: E1003 13:11:21.751692 4578 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 03 13:11:21 crc kubenswrapper[4578]: E1003 13:11:21.751799 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-etc-swift podName:aa1d7b22-8dac-42de-8241-b633bbe90691 nodeName:}" failed. No retries permitted until 2025-10-03 13:11:25.751778672 +0000 UTC m=+1221.550250856 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-etc-swift") pod "swift-storage-0" (UID: "aa1d7b22-8dac-42de-8241-b633bbe90691") : configmap "swift-ring-files" not found Oct 03 13:11:22 crc kubenswrapper[4578]: I1003 13:11:22.925243 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xhpcq" event={"ID":"35460854-45fb-4080-aad3-3638a4283374","Type":"ContainerDied","Data":"27a3ae700ec4a22e64ac474103569b546f7b11ed7a1b9a02f91ec66140a4da5a"} Oct 03 13:11:22 crc kubenswrapper[4578]: I1003 13:11:22.927009 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27a3ae700ec4a22e64ac474103569b546f7b11ed7a1b9a02f91ec66140a4da5a" Oct 03 13:11:22 crc kubenswrapper[4578]: I1003 13:11:22.950252 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xhpcq" Oct 03 13:11:22 crc kubenswrapper[4578]: I1003 13:11:22.971536 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4wbn\" (UniqueName: \"kubernetes.io/projected/35460854-45fb-4080-aad3-3638a4283374-kube-api-access-t4wbn\") pod \"35460854-45fb-4080-aad3-3638a4283374\" (UID: \"35460854-45fb-4080-aad3-3638a4283374\") " Oct 03 13:11:22 crc kubenswrapper[4578]: I1003 13:11:22.975992 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35460854-45fb-4080-aad3-3638a4283374-kube-api-access-t4wbn" (OuterVolumeSpecName: "kube-api-access-t4wbn") pod "35460854-45fb-4080-aad3-3638a4283374" (UID: "35460854-45fb-4080-aad3-3638a4283374"). InnerVolumeSpecName "kube-api-access-t4wbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:11:23 crc kubenswrapper[4578]: I1003 13:11:23.073602 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4wbn\" (UniqueName: \"kubernetes.io/projected/35460854-45fb-4080-aad3-3638a4283374-kube-api-access-t4wbn\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:23 crc kubenswrapper[4578]: I1003 13:11:23.616979 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-mz57g"] Oct 03 13:11:23 crc kubenswrapper[4578]: E1003 13:11:23.617729 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35460854-45fb-4080-aad3-3638a4283374" containerName="mariadb-database-create" Oct 03 13:11:23 crc kubenswrapper[4578]: I1003 13:11:23.617752 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="35460854-45fb-4080-aad3-3638a4283374" containerName="mariadb-database-create" Oct 03 13:11:23 crc kubenswrapper[4578]: I1003 13:11:23.617949 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="35460854-45fb-4080-aad3-3638a4283374" containerName="mariadb-database-create" Oct 03 13:11:23 crc kubenswrapper[4578]: I1003 13:11:23.618562 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mz57g" Oct 03 13:11:23 crc kubenswrapper[4578]: I1003 13:11:23.628331 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-mz57g"] Oct 03 13:11:23 crc kubenswrapper[4578]: I1003 13:11:23.709901 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmmbh\" (UniqueName: \"kubernetes.io/projected/9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc-kube-api-access-xmmbh\") pod \"keystone-db-create-mz57g\" (UID: \"9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc\") " pod="openstack/keystone-db-create-mz57g" Oct 03 13:11:23 crc kubenswrapper[4578]: I1003 13:11:23.811397 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmmbh\" (UniqueName: \"kubernetes.io/projected/9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc-kube-api-access-xmmbh\") pod \"keystone-db-create-mz57g\" (UID: \"9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc\") " pod="openstack/keystone-db-create-mz57g" Oct 03 13:11:23 crc kubenswrapper[4578]: I1003 13:11:23.828440 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmmbh\" (UniqueName: \"kubernetes.io/projected/9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc-kube-api-access-xmmbh\") pod \"keystone-db-create-mz57g\" (UID: \"9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc\") " pod="openstack/keystone-db-create-mz57g" Oct 03 13:11:23 crc kubenswrapper[4578]: I1003 13:11:23.934869 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hjk2v" event={"ID":"3f2222d5-bfbb-43ca-a40a-d235295823a8","Type":"ContainerStarted","Data":"72527ced8f2a88f2073c2b6f68d7e9e9f4a01131fa56f7a7b20af1670a9b5faa"} Oct 03 13:11:23 crc kubenswrapper[4578]: I1003 13:11:23.938952 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xhpcq" Oct 03 13:11:23 crc kubenswrapper[4578]: I1003 13:11:23.940818 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"df1fc342-37e6-4757-b1cf-81a10245997b","Type":"ContainerStarted","Data":"d3389ef2d1df96c92f4bc2fb1e178fc0a8ca5b565a32c636a4fec0e3f94e8836"} Oct 03 13:11:23 crc kubenswrapper[4578]: I1003 13:11:23.953876 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mz57g" Oct 03 13:11:23 crc kubenswrapper[4578]: I1003 13:11:23.958849 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-hjk2v" podStartSLOduration=2.041854271 podStartE2EDuration="5.958833573s" podCreationTimestamp="2025-10-03 13:11:18 +0000 UTC" firstStartedPulling="2025-10-03 13:11:19.049362478 +0000 UTC m=+1214.847834672" lastFinishedPulling="2025-10-03 13:11:22.96634179 +0000 UTC m=+1218.764813974" observedRunningTime="2025-10-03 13:11:23.9533229 +0000 UTC m=+1219.751795084" watchObservedRunningTime="2025-10-03 13:11:23.958833573 +0000 UTC m=+1219.757305757" Oct 03 13:11:23 crc kubenswrapper[4578]: I1003 13:11:23.998199 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=19.994556774 podStartE2EDuration="1m4.998183314s" podCreationTimestamp="2025-10-03 13:10:19 +0000 UTC" firstStartedPulling="2025-10-03 13:10:37.951040782 +0000 UTC m=+1173.749512966" lastFinishedPulling="2025-10-03 13:11:22.954667312 +0000 UTC m=+1218.753139506" observedRunningTime="2025-10-03 13:11:23.996022976 +0000 UTC m=+1219.794495160" watchObservedRunningTime="2025-10-03 13:11:23.998183314 +0000 UTC m=+1219.796655498" Oct 03 13:11:24 crc kubenswrapper[4578]: I1003 13:11:24.104808 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-f1e5-account-create-cvm99"] Oct 03 13:11:24 crc kubenswrapper[4578]: I1003 13:11:24.120950 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f1e5-account-create-cvm99" Oct 03 13:11:24 crc kubenswrapper[4578]: I1003 13:11:24.123849 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 03 13:11:24 crc kubenswrapper[4578]: I1003 13:11:24.129046 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f1e5-account-create-cvm99"] Oct 03 13:11:24 crc kubenswrapper[4578]: I1003 13:11:24.222269 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n68z7\" (UniqueName: \"kubernetes.io/projected/ad779b56-3577-401f-aba9-fc3b77d43200-kube-api-access-n68z7\") pod \"placement-f1e5-account-create-cvm99\" (UID: \"ad779b56-3577-401f-aba9-fc3b77d43200\") " pod="openstack/placement-f1e5-account-create-cvm99" Oct 03 13:11:24 crc kubenswrapper[4578]: I1003 13:11:24.323654 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n68z7\" (UniqueName: \"kubernetes.io/projected/ad779b56-3577-401f-aba9-fc3b77d43200-kube-api-access-n68z7\") pod \"placement-f1e5-account-create-cvm99\" (UID: \"ad779b56-3577-401f-aba9-fc3b77d43200\") " pod="openstack/placement-f1e5-account-create-cvm99" Oct 03 13:11:24 crc kubenswrapper[4578]: I1003 13:11:24.343749 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n68z7\" (UniqueName: \"kubernetes.io/projected/ad779b56-3577-401f-aba9-fc3b77d43200-kube-api-access-n68z7\") pod \"placement-f1e5-account-create-cvm99\" (UID: \"ad779b56-3577-401f-aba9-fc3b77d43200\") " pod="openstack/placement-f1e5-account-create-cvm99" Oct 03 13:11:24 crc kubenswrapper[4578]: I1003 13:11:24.409414 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-mz57g"] Oct 03 13:11:24 crc kubenswrapper[4578]: W1003 13:11:24.411552 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ee3fb9c_6fa4_4217_b262_904ebc1ef6fc.slice/crio-8187b25183f84abfc0d0bec9f57c137ac026cdd264ec3d075f749d023ba7ec02 WatchSource:0}: Error finding container 8187b25183f84abfc0d0bec9f57c137ac026cdd264ec3d075f749d023ba7ec02: Status 404 returned error can't find the container with id 8187b25183f84abfc0d0bec9f57c137ac026cdd264ec3d075f749d023ba7ec02 Oct 03 13:11:24 crc kubenswrapper[4578]: I1003 13:11:24.461017 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f1e5-account-create-cvm99" Oct 03 13:11:24 crc kubenswrapper[4578]: I1003 13:11:24.947535 4578 generic.go:334] "Generic (PLEG): container finished" podID="9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc" containerID="f46007ee12c93bed7b9fae98890c7330192ad2321fe90b5b2bf9dc8b84bb1357" exitCode=0 Oct 03 13:11:24 crc kubenswrapper[4578]: I1003 13:11:24.949320 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mz57g" event={"ID":"9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc","Type":"ContainerDied","Data":"f46007ee12c93bed7b9fae98890c7330192ad2321fe90b5b2bf9dc8b84bb1357"} Oct 03 13:11:24 crc kubenswrapper[4578]: I1003 13:11:24.949444 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mz57g" event={"ID":"9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc","Type":"ContainerStarted","Data":"8187b25183f84abfc0d0bec9f57c137ac026cdd264ec3d075f749d023ba7ec02"} Oct 03 13:11:25 crc kubenswrapper[4578]: I1003 13:11:25.018666 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f1e5-account-create-cvm99"] Oct 03 13:11:25 crc kubenswrapper[4578]: I1003 13:11:25.817055 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 03 13:11:25 crc kubenswrapper[4578]: E1003 13:11:25.849485 4578 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 03 13:11:25 crc kubenswrapper[4578]: E1003 13:11:25.849519 4578 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 03 13:11:25 crc kubenswrapper[4578]: E1003 13:11:25.849572 4578 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-etc-swift podName:aa1d7b22-8dac-42de-8241-b633bbe90691 nodeName:}" failed. No retries permitted until 2025-10-03 13:11:33.849557596 +0000 UTC m=+1229.648029780 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-etc-swift") pod "swift-storage-0" (UID: "aa1d7b22-8dac-42de-8241-b633bbe90691") : configmap "swift-ring-files" not found Oct 03 13:11:25 crc kubenswrapper[4578]: I1003 13:11:25.849594 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-etc-swift\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") " pod="openstack/swift-storage-0" Oct 03 13:11:25 crc kubenswrapper[4578]: I1003 13:11:25.957972 4578 generic.go:334] "Generic (PLEG): container finished" podID="ad779b56-3577-401f-aba9-fc3b77d43200" containerID="7a8b2e6b02891c43dc3f8afd40babd3a6d66964e50f2d9ff378202f835c0debe" exitCode=0 Oct 03 13:11:25 crc kubenswrapper[4578]: I1003 13:11:25.958018 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f1e5-account-create-cvm99" event={"ID":"ad779b56-3577-401f-aba9-fc3b77d43200","Type":"ContainerDied","Data":"7a8b2e6b02891c43dc3f8afd40babd3a6d66964e50f2d9ff378202f835c0debe"} Oct 03 13:11:25 crc kubenswrapper[4578]: I1003 13:11:25.958060 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f1e5-account-create-cvm99" event={"ID":"ad779b56-3577-401f-aba9-fc3b77d43200","Type":"ContainerStarted","Data":"32ace13af0403329d872aedbc05c0dd19c61f97221bf1278f4e569893b3f9a61"} Oct 03 13:11:26 crc kubenswrapper[4578]: I1003 13:11:26.280043 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mz57g" Oct 03 13:11:26 crc kubenswrapper[4578]: I1003 13:11:26.375916 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmmbh\" (UniqueName: \"kubernetes.io/projected/9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc-kube-api-access-xmmbh\") pod \"9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc\" (UID: \"9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc\") " Oct 03 13:11:26 crc kubenswrapper[4578]: I1003 13:11:26.383444 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc-kube-api-access-xmmbh" (OuterVolumeSpecName: "kube-api-access-xmmbh") pod "9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc" (UID: "9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc"). InnerVolumeSpecName "kube-api-access-xmmbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:11:26 crc kubenswrapper[4578]: I1003 13:11:26.477439 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmmbh\" (UniqueName: \"kubernetes.io/projected/9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc-kube-api-access-xmmbh\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:26 crc kubenswrapper[4578]: I1003 13:11:26.817539 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 03 13:11:26 crc kubenswrapper[4578]: I1003 13:11:26.865850 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 03 13:11:26 crc kubenswrapper[4578]: I1003 13:11:26.969374 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mz57g" event={"ID":"9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc","Type":"ContainerDied","Data":"8187b25183f84abfc0d0bec9f57c137ac026cdd264ec3d075f749d023ba7ec02"} Oct 03 13:11:26 crc kubenswrapper[4578]: I1003 13:11:26.969422 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8187b25183f84abfc0d0bec9f57c137ac026cdd264ec3d075f749d023ba7ec02" Oct 03 13:11:26 crc kubenswrapper[4578]: I1003 13:11:26.970306 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mz57g" Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.217747 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.274591 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f1e5-account-create-cvm99" Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.329936 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7xdrm"] Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.330200 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" podUID="2a430efe-d234-4742-af6c-40a34b16daf6" containerName="dnsmasq-dns" containerID="cri-o://0f0fa9e6c3606ed02c80167d364abd50b3bfabc42cb3579a90914845619d226b" gracePeriod=10 Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.418374 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n68z7\" (UniqueName: \"kubernetes.io/projected/ad779b56-3577-401f-aba9-fc3b77d43200-kube-api-access-n68z7\") pod \"ad779b56-3577-401f-aba9-fc3b77d43200\" (UID: \"ad779b56-3577-401f-aba9-fc3b77d43200\") " Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.423293 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad779b56-3577-401f-aba9-fc3b77d43200-kube-api-access-n68z7" (OuterVolumeSpecName: "kube-api-access-n68z7") pod "ad779b56-3577-401f-aba9-fc3b77d43200" (UID: "ad779b56-3577-401f-aba9-fc3b77d43200"). InnerVolumeSpecName "kube-api-access-n68z7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.519874 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n68z7\" (UniqueName: \"kubernetes.io/projected/ad779b56-3577-401f-aba9-fc3b77d43200-kube-api-access-n68z7\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.799044 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.926969 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a430efe-d234-4742-af6c-40a34b16daf6-config\") pod \"2a430efe-d234-4742-af6c-40a34b16daf6\" (UID: \"2a430efe-d234-4742-af6c-40a34b16daf6\") " Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.927033 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkqt6\" (UniqueName: \"kubernetes.io/projected/2a430efe-d234-4742-af6c-40a34b16daf6-kube-api-access-jkqt6\") pod \"2a430efe-d234-4742-af6c-40a34b16daf6\" (UID: \"2a430efe-d234-4742-af6c-40a34b16daf6\") " Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.927145 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a430efe-d234-4742-af6c-40a34b16daf6-dns-svc\") pod \"2a430efe-d234-4742-af6c-40a34b16daf6\" (UID: \"2a430efe-d234-4742-af6c-40a34b16daf6\") " Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.930166 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a430efe-d234-4742-af6c-40a34b16daf6-kube-api-access-jkqt6" (OuterVolumeSpecName: "kube-api-access-jkqt6") pod "2a430efe-d234-4742-af6c-40a34b16daf6" (UID: "2a430efe-d234-4742-af6c-40a34b16daf6"). InnerVolumeSpecName "kube-api-access-jkqt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.963783 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a430efe-d234-4742-af6c-40a34b16daf6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2a430efe-d234-4742-af6c-40a34b16daf6" (UID: "2a430efe-d234-4742-af6c-40a34b16daf6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.974168 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a430efe-d234-4742-af6c-40a34b16daf6-config" (OuterVolumeSpecName: "config") pod "2a430efe-d234-4742-af6c-40a34b16daf6" (UID: "2a430efe-d234-4742-af6c-40a34b16daf6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.982583 4578 generic.go:334] "Generic (PLEG): container finished" podID="2a430efe-d234-4742-af6c-40a34b16daf6" containerID="0f0fa9e6c3606ed02c80167d364abd50b3bfabc42cb3579a90914845619d226b" exitCode=0 Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.982692 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" event={"ID":"2a430efe-d234-4742-af6c-40a34b16daf6","Type":"ContainerDied","Data":"0f0fa9e6c3606ed02c80167d364abd50b3bfabc42cb3579a90914845619d226b"} Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.982726 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" event={"ID":"2a430efe-d234-4742-af6c-40a34b16daf6","Type":"ContainerDied","Data":"10fd72ed7c3aa3dffca5c9d0958c1542d926153a8597b7777b97794b7478750b"} Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.982744 4578 scope.go:117] "RemoveContainer" containerID="0f0fa9e6c3606ed02c80167d364abd50b3bfabc42cb3579a90914845619d226b" Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.982862 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7xdrm" Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.985397 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f1e5-account-create-cvm99" event={"ID":"ad779b56-3577-401f-aba9-fc3b77d43200","Type":"ContainerDied","Data":"32ace13af0403329d872aedbc05c0dd19c61f97221bf1278f4e569893b3f9a61"} Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.985437 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32ace13af0403329d872aedbc05c0dd19c61f97221bf1278f4e569893b3f9a61" Oct 03 13:11:27 crc kubenswrapper[4578]: I1003 13:11:27.985488 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f1e5-account-create-cvm99" Oct 03 13:11:28 crc kubenswrapper[4578]: I1003 13:11:28.002878 4578 scope.go:117] "RemoveContainer" containerID="d841514305e229a8be7a21be6db60c1fbefdf0a73a21295eb31c4b7636e18cbc" Oct 03 13:11:28 crc kubenswrapper[4578]: I1003 13:11:28.023108 4578 scope.go:117] "RemoveContainer" containerID="0f0fa9e6c3606ed02c80167d364abd50b3bfabc42cb3579a90914845619d226b" Oct 03 13:11:28 crc kubenswrapper[4578]: I1003 13:11:28.024494 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7xdrm"] Oct 03 13:11:28 crc kubenswrapper[4578]: E1003 13:11:28.025230 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f0fa9e6c3606ed02c80167d364abd50b3bfabc42cb3579a90914845619d226b\": container with ID starting with 0f0fa9e6c3606ed02c80167d364abd50b3bfabc42cb3579a90914845619d226b not found: ID does not exist" containerID="0f0fa9e6c3606ed02c80167d364abd50b3bfabc42cb3579a90914845619d226b" Oct 03 13:11:28 crc kubenswrapper[4578]: I1003 13:11:28.025358 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f0fa9e6c3606ed02c80167d364abd50b3bfabc42cb3579a90914845619d226b"} err="failed to get container status \"0f0fa9e6c3606ed02c80167d364abd50b3bfabc42cb3579a90914845619d226b\": rpc error: code = NotFound desc = could not find container \"0f0fa9e6c3606ed02c80167d364abd50b3bfabc42cb3579a90914845619d226b\": container with ID starting with 0f0fa9e6c3606ed02c80167d364abd50b3bfabc42cb3579a90914845619d226b not found: ID does not exist" Oct 03 13:11:28 crc kubenswrapper[4578]: I1003 13:11:28.025441 4578 scope.go:117] "RemoveContainer" containerID="d841514305e229a8be7a21be6db60c1fbefdf0a73a21295eb31c4b7636e18cbc" Oct 03 13:11:28 crc kubenswrapper[4578]: E1003 13:11:28.025885 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d841514305e229a8be7a21be6db60c1fbefdf0a73a21295eb31c4b7636e18cbc\": container with ID starting with d841514305e229a8be7a21be6db60c1fbefdf0a73a21295eb31c4b7636e18cbc not found: ID does not exist" containerID="d841514305e229a8be7a21be6db60c1fbefdf0a73a21295eb31c4b7636e18cbc" Oct 03 13:11:28 crc kubenswrapper[4578]: I1003 13:11:28.026019 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d841514305e229a8be7a21be6db60c1fbefdf0a73a21295eb31c4b7636e18cbc"} err="failed to get container status \"d841514305e229a8be7a21be6db60c1fbefdf0a73a21295eb31c4b7636e18cbc\": rpc error: code = NotFound desc = could not find container \"d841514305e229a8be7a21be6db60c1fbefdf0a73a21295eb31c4b7636e18cbc\": container with ID starting with d841514305e229a8be7a21be6db60c1fbefdf0a73a21295eb31c4b7636e18cbc not found: ID does not exist" Oct 03 13:11:28 crc kubenswrapper[4578]: I1003 13:11:28.029050 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a430efe-d234-4742-af6c-40a34b16daf6-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:28 crc kubenswrapper[4578]: I1003 13:11:28.029077 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkqt6\" (UniqueName: \"kubernetes.io/projected/2a430efe-d234-4742-af6c-40a34b16daf6-kube-api-access-jkqt6\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:28 crc kubenswrapper[4578]: I1003 13:11:28.029088 4578 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a430efe-d234-4742-af6c-40a34b16daf6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:28 crc kubenswrapper[4578]: I1003 13:11:28.032806 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7xdrm"] Oct 03 13:11:28 crc kubenswrapper[4578]: I1003 13:11:28.918490 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a430efe-d234-4742-af6c-40a34b16daf6" path="/var/lib/kubelet/pods/2a430efe-d234-4742-af6c-40a34b16daf6/volumes" Oct 03 13:11:29 crc kubenswrapper[4578]: I1003 13:11:29.563970 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-950f-account-create-br2hk"] Oct 03 13:11:29 crc kubenswrapper[4578]: E1003 13:11:29.565416 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a430efe-d234-4742-af6c-40a34b16daf6" containerName="init" Oct 03 13:11:29 crc kubenswrapper[4578]: I1003 13:11:29.565512 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a430efe-d234-4742-af6c-40a34b16daf6" containerName="init" Oct 03 13:11:29 crc kubenswrapper[4578]: E1003 13:11:29.565604 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc" containerName="mariadb-database-create" Oct 03 13:11:29 crc kubenswrapper[4578]: I1003 13:11:29.565709 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc" containerName="mariadb-database-create" Oct 03 13:11:29 crc kubenswrapper[4578]: E1003 13:11:29.565791 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a430efe-d234-4742-af6c-40a34b16daf6" containerName="dnsmasq-dns" Oct 03 13:11:29 crc kubenswrapper[4578]: I1003 13:11:29.565874 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a430efe-d234-4742-af6c-40a34b16daf6" containerName="dnsmasq-dns" Oct 03 13:11:29 crc kubenswrapper[4578]: E1003 13:11:29.566031 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad779b56-3577-401f-aba9-fc3b77d43200" containerName="mariadb-account-create" Oct 03 13:11:29 crc kubenswrapper[4578]: I1003 13:11:29.566115 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad779b56-3577-401f-aba9-fc3b77d43200" containerName="mariadb-account-create" Oct 03 13:11:29 crc kubenswrapper[4578]: I1003 13:11:29.566445 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc" containerName="mariadb-database-create" Oct 03 13:11:29 crc kubenswrapper[4578]: I1003 13:11:29.566581 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a430efe-d234-4742-af6c-40a34b16daf6" containerName="dnsmasq-dns" Oct 03 13:11:29 crc kubenswrapper[4578]: I1003 13:11:29.566957 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad779b56-3577-401f-aba9-fc3b77d43200" containerName="mariadb-account-create" Oct 03 13:11:29 crc kubenswrapper[4578]: I1003 13:11:29.567550 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-950f-account-create-br2hk" Oct 03 13:11:29 crc kubenswrapper[4578]: I1003 13:11:29.570934 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 03 13:11:29 crc kubenswrapper[4578]: I1003 13:11:29.571789 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-950f-account-create-br2hk"] Oct 03 13:11:29 crc kubenswrapper[4578]: I1003 13:11:29.656436 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfjm6\" (UniqueName: \"kubernetes.io/projected/12a533e8-e9f7-49a6-ab49-f969cf8fd70b-kube-api-access-vfjm6\") pod \"glance-950f-account-create-br2hk\" (UID: \"12a533e8-e9f7-49a6-ab49-f969cf8fd70b\") " pod="openstack/glance-950f-account-create-br2hk" Oct 03 13:11:29 crc kubenswrapper[4578]: I1003 13:11:29.757864 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfjm6\" (UniqueName: \"kubernetes.io/projected/12a533e8-e9f7-49a6-ab49-f969cf8fd70b-kube-api-access-vfjm6\") pod \"glance-950f-account-create-br2hk\" (UID: \"12a533e8-e9f7-49a6-ab49-f969cf8fd70b\") " pod="openstack/glance-950f-account-create-br2hk" Oct 03 13:11:29 crc kubenswrapper[4578]: I1003 13:11:29.776800 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfjm6\" (UniqueName: \"kubernetes.io/projected/12a533e8-e9f7-49a6-ab49-f969cf8fd70b-kube-api-access-vfjm6\") pod \"glance-950f-account-create-br2hk\" (UID: \"12a533e8-e9f7-49a6-ab49-f969cf8fd70b\") " pod="openstack/glance-950f-account-create-br2hk" Oct 03 13:11:29 crc kubenswrapper[4578]: I1003 13:11:29.887726 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-950f-account-create-br2hk" Oct 03 13:11:30 crc kubenswrapper[4578]: I1003 13:11:30.325981 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-950f-account-create-br2hk"] Oct 03 13:11:30 crc kubenswrapper[4578]: W1003 13:11:30.331567 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12a533e8_e9f7_49a6_ab49_f969cf8fd70b.slice/crio-a6ac8e113914bd02251611ddb53fa8fd98e7d387bd9b825c20114aa9531ce4d1 WatchSource:0}: Error finding container a6ac8e113914bd02251611ddb53fa8fd98e7d387bd9b825c20114aa9531ce4d1: Status 404 returned error can't find the container with id a6ac8e113914bd02251611ddb53fa8fd98e7d387bd9b825c20114aa9531ce4d1 Oct 03 13:11:30 crc kubenswrapper[4578]: I1003 13:11:30.857127 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.011802 4578 generic.go:334] "Generic (PLEG): container finished" podID="12a533e8-e9f7-49a6-ab49-f969cf8fd70b" containerID="7e7ceab89b4db745e976243499d67a364246038e2af2277eaa10066965917e54" exitCode=0 Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.011878 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-950f-account-create-br2hk" event={"ID":"12a533e8-e9f7-49a6-ab49-f969cf8fd70b","Type":"ContainerDied","Data":"7e7ceab89b4db745e976243499d67a364246038e2af2277eaa10066965917e54"} Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.011903 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-950f-account-create-br2hk" event={"ID":"12a533e8-e9f7-49a6-ab49-f969cf8fd70b","Type":"ContainerStarted","Data":"a6ac8e113914bd02251611ddb53fa8fd98e7d387bd9b825c20114aa9531ce4d1"} Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.013430 4578 generic.go:334] "Generic (PLEG): container finished" podID="3f2222d5-bfbb-43ca-a40a-d235295823a8" containerID="72527ced8f2a88f2073c2b6f68d7e9e9f4a01131fa56f7a7b20af1670a9b5faa" exitCode=0 Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.013453 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hjk2v" event={"ID":"3f2222d5-bfbb-43ca-a40a-d235295823a8","Type":"ContainerDied","Data":"72527ced8f2a88f2073c2b6f68d7e9e9f4a01131fa56f7a7b20af1670a9b5faa"} Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.104035 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.105607 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.107996 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.108377 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-r6bdm" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.108602 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.110334 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.138193 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.188524 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/54265d19-6511-40ac-9393-4cf7827def68-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.188618 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54265d19-6511-40ac-9393-4cf7827def68-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.188655 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mvj6\" (UniqueName: \"kubernetes.io/projected/54265d19-6511-40ac-9393-4cf7827def68-kube-api-access-7mvj6\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.188694 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54265d19-6511-40ac-9393-4cf7827def68-config\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.188775 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/54265d19-6511-40ac-9393-4cf7827def68-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.188795 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54265d19-6511-40ac-9393-4cf7827def68-scripts\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.188937 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/54265d19-6511-40ac-9393-4cf7827def68-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.291113 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/54265d19-6511-40ac-9393-4cf7827def68-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.291211 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/54265d19-6511-40ac-9393-4cf7827def68-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.291281 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54265d19-6511-40ac-9393-4cf7827def68-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.291308 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mvj6\" (UniqueName: \"kubernetes.io/projected/54265d19-6511-40ac-9393-4cf7827def68-kube-api-access-7mvj6\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.291352 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54265d19-6511-40ac-9393-4cf7827def68-config\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.291414 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/54265d19-6511-40ac-9393-4cf7827def68-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.291471 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54265d19-6511-40ac-9393-4cf7827def68-scripts\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.291986 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/54265d19-6511-40ac-9393-4cf7827def68-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.292401 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54265d19-6511-40ac-9393-4cf7827def68-config\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.292845 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54265d19-6511-40ac-9393-4cf7827def68-scripts\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.297461 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/54265d19-6511-40ac-9393-4cf7827def68-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.301728 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/54265d19-6511-40ac-9393-4cf7827def68-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.307684 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54265d19-6511-40ac-9393-4cf7827def68-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.310438 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mvj6\" (UniqueName: \"kubernetes.io/projected/54265d19-6511-40ac-9393-4cf7827def68-kube-api-access-7mvj6\") pod \"ovn-northd-0\" (UID: \"54265d19-6511-40ac-9393-4cf7827def68\") " pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.432964 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 03 13:11:31 crc kubenswrapper[4578]: I1003 13:11:31.863765 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 03 13:11:31 crc kubenswrapper[4578]: W1003 13:11:31.869053 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54265d19_6511_40ac_9393_4cf7827def68.slice/crio-4a044e838037e4fec2a55ba84b22bcaea8f8a713db08876063507fad91369d11 WatchSource:0}: Error finding container 4a044e838037e4fec2a55ba84b22bcaea8f8a713db08876063507fad91369d11: Status 404 returned error can't find the container with id 4a044e838037e4fec2a55ba84b22bcaea8f8a713db08876063507fad91369d11 Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.020848 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"54265d19-6511-40ac-9393-4cf7827def68","Type":"ContainerStarted","Data":"4a044e838037e4fec2a55ba84b22bcaea8f8a713db08876063507fad91369d11"} Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.419412 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.424853 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-950f-account-create-br2hk" Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.509616 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3f2222d5-bfbb-43ca-a40a-d235295823a8-ring-data-devices\") pod \"3f2222d5-bfbb-43ca-a40a-d235295823a8\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.509721 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3f2222d5-bfbb-43ca-a40a-d235295823a8-dispersionconf\") pod \"3f2222d5-bfbb-43ca-a40a-d235295823a8\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.509805 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f2222d5-bfbb-43ca-a40a-d235295823a8-scripts\") pod \"3f2222d5-bfbb-43ca-a40a-d235295823a8\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.509834 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zmfn\" (UniqueName: \"kubernetes.io/projected/3f2222d5-bfbb-43ca-a40a-d235295823a8-kube-api-access-4zmfn\") pod \"3f2222d5-bfbb-43ca-a40a-d235295823a8\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.509863 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3f2222d5-bfbb-43ca-a40a-d235295823a8-etc-swift\") pod \"3f2222d5-bfbb-43ca-a40a-d235295823a8\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.509951 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2222d5-bfbb-43ca-a40a-d235295823a8-combined-ca-bundle\") pod \"3f2222d5-bfbb-43ca-a40a-d235295823a8\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.509988 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfjm6\" (UniqueName: \"kubernetes.io/projected/12a533e8-e9f7-49a6-ab49-f969cf8fd70b-kube-api-access-vfjm6\") pod \"12a533e8-e9f7-49a6-ab49-f969cf8fd70b\" (UID: \"12a533e8-e9f7-49a6-ab49-f969cf8fd70b\") " Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.510010 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3f2222d5-bfbb-43ca-a40a-d235295823a8-swiftconf\") pod \"3f2222d5-bfbb-43ca-a40a-d235295823a8\" (UID: \"3f2222d5-bfbb-43ca-a40a-d235295823a8\") " Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.510469 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f2222d5-bfbb-43ca-a40a-d235295823a8-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3f2222d5-bfbb-43ca-a40a-d235295823a8" (UID: "3f2222d5-bfbb-43ca-a40a-d235295823a8"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.511112 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f2222d5-bfbb-43ca-a40a-d235295823a8-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3f2222d5-bfbb-43ca-a40a-d235295823a8" (UID: "3f2222d5-bfbb-43ca-a40a-d235295823a8"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.519264 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2222d5-bfbb-43ca-a40a-d235295823a8-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3f2222d5-bfbb-43ca-a40a-d235295823a8" (UID: "3f2222d5-bfbb-43ca-a40a-d235295823a8"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.521594 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12a533e8-e9f7-49a6-ab49-f969cf8fd70b-kube-api-access-vfjm6" (OuterVolumeSpecName: "kube-api-access-vfjm6") pod "12a533e8-e9f7-49a6-ab49-f969cf8fd70b" (UID: "12a533e8-e9f7-49a6-ab49-f969cf8fd70b"). InnerVolumeSpecName "kube-api-access-vfjm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.532514 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f2222d5-bfbb-43ca-a40a-d235295823a8-kube-api-access-4zmfn" (OuterVolumeSpecName: "kube-api-access-4zmfn") pod "3f2222d5-bfbb-43ca-a40a-d235295823a8" (UID: "3f2222d5-bfbb-43ca-a40a-d235295823a8"). InnerVolumeSpecName "kube-api-access-4zmfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.533083 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2222d5-bfbb-43ca-a40a-d235295823a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f2222d5-bfbb-43ca-a40a-d235295823a8" (UID: "3f2222d5-bfbb-43ca-a40a-d235295823a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.533696 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f2222d5-bfbb-43ca-a40a-d235295823a8-scripts" (OuterVolumeSpecName: "scripts") pod "3f2222d5-bfbb-43ca-a40a-d235295823a8" (UID: "3f2222d5-bfbb-43ca-a40a-d235295823a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.538142 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2222d5-bfbb-43ca-a40a-d235295823a8-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3f2222d5-bfbb-43ca-a40a-d235295823a8" (UID: "3f2222d5-bfbb-43ca-a40a-d235295823a8"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.612144 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2222d5-bfbb-43ca-a40a-d235295823a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.612644 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfjm6\" (UniqueName: \"kubernetes.io/projected/12a533e8-e9f7-49a6-ab49-f969cf8fd70b-kube-api-access-vfjm6\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.612664 4578 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3f2222d5-bfbb-43ca-a40a-d235295823a8-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.612678 4578 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3f2222d5-bfbb-43ca-a40a-d235295823a8-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.612688 4578 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3f2222d5-bfbb-43ca-a40a-d235295823a8-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.612698 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f2222d5-bfbb-43ca-a40a-d235295823a8-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.612708 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zmfn\" (UniqueName: \"kubernetes.io/projected/3f2222d5-bfbb-43ca-a40a-d235295823a8-kube-api-access-4zmfn\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:32 crc kubenswrapper[4578]: I1003 13:11:32.612721 4578 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3f2222d5-bfbb-43ca-a40a-d235295823a8-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.033821 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-hjk2v" event={"ID":"3f2222d5-bfbb-43ca-a40a-d235295823a8","Type":"ContainerDied","Data":"9d3f4c700f8e6974ef2e67b81b68e3121b8f7048ffc8fa93dfafbe27917ee666"} Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.033860 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hjk2v" Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.033869 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d3f4c700f8e6974ef2e67b81b68e3121b8f7048ffc8fa93dfafbe27917ee666" Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.035362 4578 generic.go:334] "Generic (PLEG): container finished" podID="79fcccfa-6df4-41f2-965b-357e5a7984ab" containerID="ff3fb681d8b8aa234547e28273c535c33bdeabb40fca2a2d071b337660eef7cd" exitCode=0 Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.035435 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"79fcccfa-6df4-41f2-965b-357e5a7984ab","Type":"ContainerDied","Data":"ff3fb681d8b8aa234547e28273c535c33bdeabb40fca2a2d071b337660eef7cd"} Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.041063 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-950f-account-create-br2hk" event={"ID":"12a533e8-e9f7-49a6-ab49-f969cf8fd70b","Type":"ContainerDied","Data":"a6ac8e113914bd02251611ddb53fa8fd98e7d387bd9b825c20114aa9531ce4d1"} Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.041139 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6ac8e113914bd02251611ddb53fa8fd98e7d387bd9b825c20114aa9531ce4d1" Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.041227 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-950f-account-create-br2hk" Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.057082 4578 generic.go:334] "Generic (PLEG): container finished" podID="c196bfb1-793e-4e04-9602-db06886385ad" containerID="d3c16cfe0b8487b875181d1c37dc0cd7130f888f36465be2bb67dafa2f43ecaa" exitCode=0 Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.057132 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c196bfb1-793e-4e04-9602-db06886385ad","Type":"ContainerDied","Data":"d3c16cfe0b8487b875181d1c37dc0cd7130f888f36465be2bb67dafa2f43ecaa"} Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.737411 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-538c-account-create-kk4z8"] Oct 03 13:11:33 crc kubenswrapper[4578]: E1003 13:11:33.742082 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12a533e8-e9f7-49a6-ab49-f969cf8fd70b" containerName="mariadb-account-create" Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.742107 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="12a533e8-e9f7-49a6-ab49-f969cf8fd70b" containerName="mariadb-account-create" Oct 03 13:11:33 crc kubenswrapper[4578]: E1003 13:11:33.742176 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f2222d5-bfbb-43ca-a40a-d235295823a8" containerName="swift-ring-rebalance" Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.742184 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f2222d5-bfbb-43ca-a40a-d235295823a8" containerName="swift-ring-rebalance" Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.742322 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f2222d5-bfbb-43ca-a40a-d235295823a8" containerName="swift-ring-rebalance" Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.742339 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="12a533e8-e9f7-49a6-ab49-f969cf8fd70b" containerName="mariadb-account-create" Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.742894 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-538c-account-create-kk4z8" Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.745409 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.756597 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-538c-account-create-kk4z8"] Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.848172 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49tzw\" (UniqueName: \"kubernetes.io/projected/618c7600-498d-48ba-afd8-5b9d81638236-kube-api-access-49tzw\") pod \"keystone-538c-account-create-kk4z8\" (UID: \"618c7600-498d-48ba-afd8-5b9d81638236\") " pod="openstack/keystone-538c-account-create-kk4z8" Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.950074 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-etc-swift\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") " pod="openstack/swift-storage-0" Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.950158 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49tzw\" (UniqueName: \"kubernetes.io/projected/618c7600-498d-48ba-afd8-5b9d81638236-kube-api-access-49tzw\") pod \"keystone-538c-account-create-kk4z8\" (UID: \"618c7600-498d-48ba-afd8-5b9d81638236\") " pod="openstack/keystone-538c-account-create-kk4z8" Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.957528 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/aa1d7b22-8dac-42de-8241-b633bbe90691-etc-swift\") pod \"swift-storage-0\" (UID: \"aa1d7b22-8dac-42de-8241-b633bbe90691\") " pod="openstack/swift-storage-0" Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.972173 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.976292 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49tzw\" (UniqueName: \"kubernetes.io/projected/618c7600-498d-48ba-afd8-5b9d81638236-kube-api-access-49tzw\") pod \"keystone-538c-account-create-kk4z8\" (UID: \"618c7600-498d-48ba-afd8-5b9d81638236\") " pod="openstack/keystone-538c-account-create-kk4z8" Oct 03 13:11:33 crc kubenswrapper[4578]: I1003 13:11:33.992277 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-2sqbb" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.066466 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"54265d19-6511-40ac-9393-4cf7827def68","Type":"ContainerStarted","Data":"3810eb141d86e47c70688a88ee781cc2db5de927731bcde87d84face75800360"} Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.067615 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.067755 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"54265d19-6511-40ac-9393-4cf7827def68","Type":"ContainerStarted","Data":"81cfe7f0ef1621148b1359f98342661dbe9f65be06f26c34112e88e37ee09b58"} Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.069911 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"79fcccfa-6df4-41f2-965b-357e5a7984ab","Type":"ContainerStarted","Data":"6c285b7d81188b6d06e8bd70cfd27aeb562c9da338b801699ea591e8e5889ee9"} Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.070164 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.072053 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c196bfb1-793e-4e04-9602-db06886385ad","Type":"ContainerStarted","Data":"41aac4bc65fb0d271edee3e21b8ce5e0add0172a06237301735746d616fb09e3"} Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.071903 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-538c-account-create-kk4z8" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.073026 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.090098 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.737344389 podStartE2EDuration="3.090079034s" podCreationTimestamp="2025-10-03 13:11:31 +0000 UTC" firstStartedPulling="2025-10-03 13:11:31.871387106 +0000 UTC m=+1227.669859290" lastFinishedPulling="2025-10-03 13:11:33.224121751 +0000 UTC m=+1229.022593935" observedRunningTime="2025-10-03 13:11:34.089203326 +0000 UTC m=+1229.887675510" watchObservedRunningTime="2025-10-03 13:11:34.090079034 +0000 UTC m=+1229.888551218" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.120229 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.070647579 podStartE2EDuration="1m25.120211254s" podCreationTimestamp="2025-10-03 13:10:09 +0000 UTC" firstStartedPulling="2025-10-03 13:10:11.085540116 +0000 UTC m=+1146.884012300" lastFinishedPulling="2025-10-03 13:11:00.135103791 +0000 UTC m=+1195.933575975" observedRunningTime="2025-10-03 13:11:34.118667785 +0000 UTC m=+1229.917139959" watchObservedRunningTime="2025-10-03 13:11:34.120211254 +0000 UTC m=+1229.918683438" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.153188 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.730321968 podStartE2EDuration="1m26.153164813s" podCreationTimestamp="2025-10-03 13:10:08 +0000 UTC" firstStartedPulling="2025-10-03 13:10:10.711298456 +0000 UTC m=+1146.509770630" lastFinishedPulling="2025-10-03 13:11:00.134141291 +0000 UTC m=+1195.932613475" observedRunningTime="2025-10-03 13:11:34.150645744 +0000 UTC m=+1229.949117928" watchObservedRunningTime="2025-10-03 13:11:34.153164813 +0000 UTC m=+1229.951636997" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.232764 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-9jl4l-config-mjwh4"] Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.236830 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.265460 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.265620 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.277164 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9jl4l-config-mjwh4"] Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.298584 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-9jl4l" podUID="daabdc14-9eaa-478d-8e0f-3ab92c0568ce" containerName="ovn-controller" probeResult="failure" output=< Oct 03 13:11:34 crc kubenswrapper[4578]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 03 13:11:34 crc kubenswrapper[4578]: > Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.367051 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-additional-scripts\") pod \"ovn-controller-9jl4l-config-mjwh4\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.367683 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn4gr\" (UniqueName: \"kubernetes.io/projected/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-kube-api-access-hn4gr\") pod \"ovn-controller-9jl4l-config-mjwh4\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.368103 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-var-log-ovn\") pod \"ovn-controller-9jl4l-config-mjwh4\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.368318 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-var-run-ovn\") pod \"ovn-controller-9jl4l-config-mjwh4\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.368352 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-var-run\") pod \"ovn-controller-9jl4l-config-mjwh4\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.368451 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-scripts\") pod \"ovn-controller-9jl4l-config-mjwh4\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.471972 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-scripts\") pod \"ovn-controller-9jl4l-config-mjwh4\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.472045 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-additional-scripts\") pod \"ovn-controller-9jl4l-config-mjwh4\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.472097 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn4gr\" (UniqueName: \"kubernetes.io/projected/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-kube-api-access-hn4gr\") pod \"ovn-controller-9jl4l-config-mjwh4\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.472169 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-var-log-ovn\") pod \"ovn-controller-9jl4l-config-mjwh4\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.472205 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-var-run-ovn\") pod \"ovn-controller-9jl4l-config-mjwh4\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.472229 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-var-run\") pod \"ovn-controller-9jl4l-config-mjwh4\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.472897 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-var-run\") pod \"ovn-controller-9jl4l-config-mjwh4\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.472936 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-var-log-ovn\") pod \"ovn-controller-9jl4l-config-mjwh4\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.472956 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-var-run-ovn\") pod \"ovn-controller-9jl4l-config-mjwh4\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.473556 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-additional-scripts\") pod \"ovn-controller-9jl4l-config-mjwh4\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.474438 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-scripts\") pod \"ovn-controller-9jl4l-config-mjwh4\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.494551 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn4gr\" (UniqueName: \"kubernetes.io/projected/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-kube-api-access-hn4gr\") pod \"ovn-controller-9jl4l-config-mjwh4\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.591325 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.674463 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-538c-account-create-kk4z8"] Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.751113 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-x6xhq"] Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.752028 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-x6xhq" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.759214 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-54nlz" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.759392 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.779576 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-x6xhq"] Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.879068 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7542b422-3e5f-479c-ae79-d3209197d088-combined-ca-bundle\") pod \"glance-db-sync-x6xhq\" (UID: \"7542b422-3e5f-479c-ae79-d3209197d088\") " pod="openstack/glance-db-sync-x6xhq" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.879377 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nl84s\" (UniqueName: \"kubernetes.io/projected/7542b422-3e5f-479c-ae79-d3209197d088-kube-api-access-nl84s\") pod \"glance-db-sync-x6xhq\" (UID: \"7542b422-3e5f-479c-ae79-d3209197d088\") " pod="openstack/glance-db-sync-x6xhq" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.879407 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7542b422-3e5f-479c-ae79-d3209197d088-db-sync-config-data\") pod \"glance-db-sync-x6xhq\" (UID: \"7542b422-3e5f-479c-ae79-d3209197d088\") " pod="openstack/glance-db-sync-x6xhq" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.879428 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7542b422-3e5f-479c-ae79-d3209197d088-config-data\") pod \"glance-db-sync-x6xhq\" (UID: \"7542b422-3e5f-479c-ae79-d3209197d088\") " pod="openstack/glance-db-sync-x6xhq" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.983458 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nl84s\" (UniqueName: \"kubernetes.io/projected/7542b422-3e5f-479c-ae79-d3209197d088-kube-api-access-nl84s\") pod \"glance-db-sync-x6xhq\" (UID: \"7542b422-3e5f-479c-ae79-d3209197d088\") " pod="openstack/glance-db-sync-x6xhq" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.983763 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7542b422-3e5f-479c-ae79-d3209197d088-db-sync-config-data\") pod \"glance-db-sync-x6xhq\" (UID: \"7542b422-3e5f-479c-ae79-d3209197d088\") " pod="openstack/glance-db-sync-x6xhq" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.983888 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7542b422-3e5f-479c-ae79-d3209197d088-config-data\") pod \"glance-db-sync-x6xhq\" (UID: \"7542b422-3e5f-479c-ae79-d3209197d088\") " pod="openstack/glance-db-sync-x6xhq" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.984036 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7542b422-3e5f-479c-ae79-d3209197d088-combined-ca-bundle\") pod \"glance-db-sync-x6xhq\" (UID: \"7542b422-3e5f-479c-ae79-d3209197d088\") " pod="openstack/glance-db-sync-x6xhq" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.986365 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.988609 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7542b422-3e5f-479c-ae79-d3209197d088-config-data\") pod \"glance-db-sync-x6xhq\" (UID: \"7542b422-3e5f-479c-ae79-d3209197d088\") " pod="openstack/glance-db-sync-x6xhq" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.989020 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7542b422-3e5f-479c-ae79-d3209197d088-db-sync-config-data\") pod \"glance-db-sync-x6xhq\" (UID: \"7542b422-3e5f-479c-ae79-d3209197d088\") " pod="openstack/glance-db-sync-x6xhq" Oct 03 13:11:34 crc kubenswrapper[4578]: I1003 13:11:34.991844 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7542b422-3e5f-479c-ae79-d3209197d088-combined-ca-bundle\") pod \"glance-db-sync-x6xhq\" (UID: \"7542b422-3e5f-479c-ae79-d3209197d088\") " pod="openstack/glance-db-sync-x6xhq" Oct 03 13:11:35 crc kubenswrapper[4578]: I1003 13:11:35.018968 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nl84s\" (UniqueName: \"kubernetes.io/projected/7542b422-3e5f-479c-ae79-d3209197d088-kube-api-access-nl84s\") pod \"glance-db-sync-x6xhq\" (UID: \"7542b422-3e5f-479c-ae79-d3209197d088\") " pod="openstack/glance-db-sync-x6xhq" Oct 03 13:11:35 crc kubenswrapper[4578]: I1003 13:11:35.077908 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aa1d7b22-8dac-42de-8241-b633bbe90691","Type":"ContainerStarted","Data":"d21d3222cc52fc0e5f11f6b1ec36bcf809f5125bd9e14cd69997dd8302841363"} Oct 03 13:11:35 crc kubenswrapper[4578]: I1003 13:11:35.079108 4578 generic.go:334] "Generic (PLEG): container finished" podID="618c7600-498d-48ba-afd8-5b9d81638236" containerID="5dc6396f7f18ad1c0394a563d4a36cda87300d00170c0f6c07959c20c6032a6d" exitCode=0 Oct 03 13:11:35 crc kubenswrapper[4578]: I1003 13:11:35.079249 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-538c-account-create-kk4z8" event={"ID":"618c7600-498d-48ba-afd8-5b9d81638236","Type":"ContainerDied","Data":"5dc6396f7f18ad1c0394a563d4a36cda87300d00170c0f6c07959c20c6032a6d"} Oct 03 13:11:35 crc kubenswrapper[4578]: I1003 13:11:35.079272 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-538c-account-create-kk4z8" event={"ID":"618c7600-498d-48ba-afd8-5b9d81638236","Type":"ContainerStarted","Data":"81cfa80f3be3c0ad258b2c8885a2168619a550166f17d2f57ae21af69439ac48"} Oct 03 13:11:35 crc kubenswrapper[4578]: I1003 13:11:35.080902 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-x6xhq" Oct 03 13:11:35 crc kubenswrapper[4578]: I1003 13:11:35.091376 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:11:35 crc kubenswrapper[4578]: I1003 13:11:35.091430 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:11:35 crc kubenswrapper[4578]: I1003 13:11:35.252407 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9jl4l-config-mjwh4"] Oct 03 13:11:35 crc kubenswrapper[4578]: I1003 13:11:35.621405 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-x6xhq"] Oct 03 13:11:35 crc kubenswrapper[4578]: W1003 13:11:35.630899 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7542b422_3e5f_479c_ae79_d3209197d088.slice/crio-2734b6120170924d6cdd842bba117d13d842570c3725ce053a5c8bc85fc09f19 WatchSource:0}: Error finding container 2734b6120170924d6cdd842bba117d13d842570c3725ce053a5c8bc85fc09f19: Status 404 returned error can't find the container with id 2734b6120170924d6cdd842bba117d13d842570c3725ce053a5c8bc85fc09f19 Oct 03 13:11:36 crc kubenswrapper[4578]: I1003 13:11:36.097572 4578 generic.go:334] "Generic (PLEG): container finished" podID="b9ae88e4-98b7-425f-98c2-2cb77be46b1e" containerID="908f345453ffb6da3b63f9535168cdc1bdc29491e2e16489faccc06fb38326f0" exitCode=0 Oct 03 13:11:36 crc kubenswrapper[4578]: I1003 13:11:36.097775 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9jl4l-config-mjwh4" event={"ID":"b9ae88e4-98b7-425f-98c2-2cb77be46b1e","Type":"ContainerDied","Data":"908f345453ffb6da3b63f9535168cdc1bdc29491e2e16489faccc06fb38326f0"} Oct 03 13:11:36 crc kubenswrapper[4578]: I1003 13:11:36.097934 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9jl4l-config-mjwh4" event={"ID":"b9ae88e4-98b7-425f-98c2-2cb77be46b1e","Type":"ContainerStarted","Data":"c3abe543f254dc10252ffc09d31bd118cb71ad4f5ae5105b53572bc4ea6bc21a"} Oct 03 13:11:36 crc kubenswrapper[4578]: I1003 13:11:36.103508 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-x6xhq" event={"ID":"7542b422-3e5f-479c-ae79-d3209197d088","Type":"ContainerStarted","Data":"2734b6120170924d6cdd842bba117d13d842570c3725ce053a5c8bc85fc09f19"} Oct 03 13:11:36 crc kubenswrapper[4578]: I1003 13:11:36.972561 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-538c-account-create-kk4z8" Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.016050 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49tzw\" (UniqueName: \"kubernetes.io/projected/618c7600-498d-48ba-afd8-5b9d81638236-kube-api-access-49tzw\") pod \"618c7600-498d-48ba-afd8-5b9d81638236\" (UID: \"618c7600-498d-48ba-afd8-5b9d81638236\") " Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.030856 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/618c7600-498d-48ba-afd8-5b9d81638236-kube-api-access-49tzw" (OuterVolumeSpecName: "kube-api-access-49tzw") pod "618c7600-498d-48ba-afd8-5b9d81638236" (UID: "618c7600-498d-48ba-afd8-5b9d81638236"). InnerVolumeSpecName "kube-api-access-49tzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.112505 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-538c-account-create-kk4z8" Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.112724 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-538c-account-create-kk4z8" event={"ID":"618c7600-498d-48ba-afd8-5b9d81638236","Type":"ContainerDied","Data":"81cfa80f3be3c0ad258b2c8885a2168619a550166f17d2f57ae21af69439ac48"} Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.112759 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81cfa80f3be3c0ad258b2c8885a2168619a550166f17d2f57ae21af69439ac48" Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.118748 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49tzw\" (UniqueName: \"kubernetes.io/projected/618c7600-498d-48ba-afd8-5b9d81638236-kube-api-access-49tzw\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.365557 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.427069 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-var-run-ovn\") pod \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.427166 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-scripts\") pod \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.427209 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-additional-scripts\") pod \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.427241 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hn4gr\" (UniqueName: \"kubernetes.io/projected/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-kube-api-access-hn4gr\") pod \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.427297 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-var-log-ovn\") pod \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.427397 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-var-run\") pod \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\" (UID: \"b9ae88e4-98b7-425f-98c2-2cb77be46b1e\") " Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.428039 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-var-run" (OuterVolumeSpecName: "var-run") pod "b9ae88e4-98b7-425f-98c2-2cb77be46b1e" (UID: "b9ae88e4-98b7-425f-98c2-2cb77be46b1e"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.428082 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "b9ae88e4-98b7-425f-98c2-2cb77be46b1e" (UID: "b9ae88e4-98b7-425f-98c2-2cb77be46b1e"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.429482 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-scripts" (OuterVolumeSpecName: "scripts") pod "b9ae88e4-98b7-425f-98c2-2cb77be46b1e" (UID: "b9ae88e4-98b7-425f-98c2-2cb77be46b1e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.429781 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "b9ae88e4-98b7-425f-98c2-2cb77be46b1e" (UID: "b9ae88e4-98b7-425f-98c2-2cb77be46b1e"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.431126 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "b9ae88e4-98b7-425f-98c2-2cb77be46b1e" (UID: "b9ae88e4-98b7-425f-98c2-2cb77be46b1e"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.437904 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-kube-api-access-hn4gr" (OuterVolumeSpecName: "kube-api-access-hn4gr") pod "b9ae88e4-98b7-425f-98c2-2cb77be46b1e" (UID: "b9ae88e4-98b7-425f-98c2-2cb77be46b1e"). InnerVolumeSpecName "kube-api-access-hn4gr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.529340 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hn4gr\" (UniqueName: \"kubernetes.io/projected/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-kube-api-access-hn4gr\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.529368 4578 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.529380 4578 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-var-run\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.529390 4578 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.529399 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:37 crc kubenswrapper[4578]: I1003 13:11:37.529406 4578 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b9ae88e4-98b7-425f-98c2-2cb77be46b1e-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:38 crc kubenswrapper[4578]: I1003 13:11:38.123140 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aa1d7b22-8dac-42de-8241-b633bbe90691","Type":"ContainerStarted","Data":"3ca4c56dbc46afa7c236b2ed0c5e0180154dc391f1e456024b858a4c9e4f87ae"} Oct 03 13:11:38 crc kubenswrapper[4578]: I1003 13:11:38.123183 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aa1d7b22-8dac-42de-8241-b633bbe90691","Type":"ContainerStarted","Data":"054532f8403f34d52348bd2554a48335ee9df8faf7d0bef6c049ce9654c46dd1"} Oct 03 13:11:38 crc kubenswrapper[4578]: I1003 13:11:38.123193 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aa1d7b22-8dac-42de-8241-b633bbe90691","Type":"ContainerStarted","Data":"bff5f073d9afd6652ee48fcce61428dad3945f52e46f9f9b88e50b720bff0621"} Oct 03 13:11:38 crc kubenswrapper[4578]: I1003 13:11:38.123201 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aa1d7b22-8dac-42de-8241-b633bbe90691","Type":"ContainerStarted","Data":"bbce189de1e4a16cc86820ff92d869d5cb64689cad0e843c3dfea181e209565b"} Oct 03 13:11:38 crc kubenswrapper[4578]: I1003 13:11:38.125736 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9jl4l-config-mjwh4" event={"ID":"b9ae88e4-98b7-425f-98c2-2cb77be46b1e","Type":"ContainerDied","Data":"c3abe543f254dc10252ffc09d31bd118cb71ad4f5ae5105b53572bc4ea6bc21a"} Oct 03 13:11:38 crc kubenswrapper[4578]: I1003 13:11:38.125792 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9jl4l-config-mjwh4" Oct 03 13:11:38 crc kubenswrapper[4578]: I1003 13:11:38.125798 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3abe543f254dc10252ffc09d31bd118cb71ad4f5ae5105b53572bc4ea6bc21a" Oct 03 13:11:38 crc kubenswrapper[4578]: I1003 13:11:38.507371 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-9jl4l-config-mjwh4"] Oct 03 13:11:38 crc kubenswrapper[4578]: I1003 13:11:38.518163 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-9jl4l-config-mjwh4"] Oct 03 13:11:38 crc kubenswrapper[4578]: I1003 13:11:38.923977 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9ae88e4-98b7-425f-98c2-2cb77be46b1e" path="/var/lib/kubelet/pods/b9ae88e4-98b7-425f-98c2-2cb77be46b1e/volumes" Oct 03 13:11:39 crc kubenswrapper[4578]: I1003 13:11:39.247229 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-9jl4l" Oct 03 13:11:40 crc kubenswrapper[4578]: I1003 13:11:40.147752 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aa1d7b22-8dac-42de-8241-b633bbe90691","Type":"ContainerStarted","Data":"a015633cb96b85e52376e6c3ae71753acaca092eaf9e8c1cf62b4f78699a6b59"} Oct 03 13:11:40 crc kubenswrapper[4578]: I1003 13:11:40.148104 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aa1d7b22-8dac-42de-8241-b633bbe90691","Type":"ContainerStarted","Data":"c0765bb9618dd20f4020a1ce5ee731056641ecc4e5b55c7802ca854fbd90ecc3"} Oct 03 13:11:40 crc kubenswrapper[4578]: I1003 13:11:40.148121 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aa1d7b22-8dac-42de-8241-b633bbe90691","Type":"ContainerStarted","Data":"f735ba548997ce8a1754781138e72dee5387ac146f381161838ea993b68c6ec9"} Oct 03 13:11:41 crc kubenswrapper[4578]: I1003 13:11:41.161761 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aa1d7b22-8dac-42de-8241-b633bbe90691","Type":"ContainerStarted","Data":"0893d4aabaff683c51040826ab2a05bf78be066749432c9458bd4e572e15f944"} Oct 03 13:11:43 crc kubenswrapper[4578]: I1003 13:11:43.180115 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aa1d7b22-8dac-42de-8241-b633bbe90691","Type":"ContainerStarted","Data":"f2729473dd97881f52886dbdedda13cfd22d07c70b174dc973864dbb8d78e6b7"} Oct 03 13:11:43 crc kubenswrapper[4578]: I1003 13:11:43.180552 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aa1d7b22-8dac-42de-8241-b633bbe90691","Type":"ContainerStarted","Data":"bf0c64a1b2f96340885d4033103acc3dae6cf7ba90864c7beeb212f2f3a40caa"} Oct 03 13:11:43 crc kubenswrapper[4578]: I1003 13:11:43.180571 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aa1d7b22-8dac-42de-8241-b633bbe90691","Type":"ContainerStarted","Data":"33b521de0a2becbf9060713e90ad279aa796ecb326fe98596633b2923f88ee55"} Oct 03 13:11:46 crc kubenswrapper[4578]: I1003 13:11:46.561361 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.061737 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.305604 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aa1d7b22-8dac-42de-8241-b633bbe90691","Type":"ContainerStarted","Data":"a3f2e0f3117370d7464bb07d817d51d0a897b2005fff2ebfaf75a6feb9b7da8c"} Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.305988 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aa1d7b22-8dac-42de-8241-b633bbe90691","Type":"ContainerStarted","Data":"06f572dfb77e41d99311493bafd96d65adc4339e25e036cb5a5a243c9e7bb56d"} Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.306002 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aa1d7b22-8dac-42de-8241-b633bbe90691","Type":"ContainerStarted","Data":"3ed97cc2c5037556dde6793c91daf23ab54f76cc290c6d87fdabdde39fc825f9"} Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.389263 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-twh27"] Oct 03 13:11:50 crc kubenswrapper[4578]: E1003 13:11:50.389577 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="618c7600-498d-48ba-afd8-5b9d81638236" containerName="mariadb-account-create" Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.389595 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="618c7600-498d-48ba-afd8-5b9d81638236" containerName="mariadb-account-create" Oct 03 13:11:50 crc kubenswrapper[4578]: E1003 13:11:50.389653 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9ae88e4-98b7-425f-98c2-2cb77be46b1e" containerName="ovn-config" Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.389660 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9ae88e4-98b7-425f-98c2-2cb77be46b1e" containerName="ovn-config" Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.389833 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9ae88e4-98b7-425f-98c2-2cb77be46b1e" containerName="ovn-config" Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.389863 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="618c7600-498d-48ba-afd8-5b9d81638236" containerName="mariadb-account-create" Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.390393 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-twh27" Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.412469 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-twh27"] Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.453498 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhqzm\" (UniqueName: \"kubernetes.io/projected/8999650e-96de-4898-9589-156a996042ed-kube-api-access-qhqzm\") pod \"cinder-db-create-twh27\" (UID: \"8999650e-96de-4898-9589-156a996042ed\") " pod="openstack/cinder-db-create-twh27" Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.467798 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.566438 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhqzm\" (UniqueName: \"kubernetes.io/projected/8999650e-96de-4898-9589-156a996042ed-kube-api-access-qhqzm\") pod \"cinder-db-create-twh27\" (UID: \"8999650e-96de-4898-9589-156a996042ed\") " pod="openstack/cinder-db-create-twh27" Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.611097 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhqzm\" (UniqueName: \"kubernetes.io/projected/8999650e-96de-4898-9589-156a996042ed-kube-api-access-qhqzm\") pod \"cinder-db-create-twh27\" (UID: \"8999650e-96de-4898-9589-156a996042ed\") " pod="openstack/cinder-db-create-twh27" Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.654579 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-vzh7p"] Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.655866 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-vzh7p" Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.682648 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-vzh7p"] Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.706073 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-twh27" Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.772685 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcrzl\" (UniqueName: \"kubernetes.io/projected/1953665a-f20b-4620-8061-bc3e4a33aacd-kube-api-access-xcrzl\") pod \"barbican-db-create-vzh7p\" (UID: \"1953665a-f20b-4620-8061-bc3e4a33aacd\") " pod="openstack/barbican-db-create-vzh7p" Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.875539 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcrzl\" (UniqueName: \"kubernetes.io/projected/1953665a-f20b-4620-8061-bc3e4a33aacd-kube-api-access-xcrzl\") pod \"barbican-db-create-vzh7p\" (UID: \"1953665a-f20b-4620-8061-bc3e4a33aacd\") " pod="openstack/barbican-db-create-vzh7p" Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.900325 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcrzl\" (UniqueName: \"kubernetes.io/projected/1953665a-f20b-4620-8061-bc3e4a33aacd-kube-api-access-xcrzl\") pod \"barbican-db-create-vzh7p\" (UID: \"1953665a-f20b-4620-8061-bc3e4a33aacd\") " pod="openstack/barbican-db-create-vzh7p" Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.981244 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-vzh7p" Oct 03 13:11:50 crc kubenswrapper[4578]: I1003 13:11:50.992868 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-wsqj9"] Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.004202 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-wsqj9" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.028779 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-wsqj9"] Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.082287 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnjj8\" (UniqueName: \"kubernetes.io/projected/beb0dafe-f425-4969-b586-29ca36b936f0-kube-api-access-gnjj8\") pod \"neutron-db-create-wsqj9\" (UID: \"beb0dafe-f425-4969-b586-29ca36b936f0\") " pod="openstack/neutron-db-create-wsqj9" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.092901 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-9znjc"] Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.095915 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-9znjc" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.098023 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wxw7h" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.099403 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.099532 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.107391 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.134880 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-9znjc"] Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.184158 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca2b580-f2b5-4f61-9aa8-63b114198bb0-combined-ca-bundle\") pod \"keystone-db-sync-9znjc\" (UID: \"6ca2b580-f2b5-4f61-9aa8-63b114198bb0\") " pod="openstack/keystone-db-sync-9znjc" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.184280 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ca2b580-f2b5-4f61-9aa8-63b114198bb0-config-data\") pod \"keystone-db-sync-9znjc\" (UID: \"6ca2b580-f2b5-4f61-9aa8-63b114198bb0\") " pod="openstack/keystone-db-sync-9znjc" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.184314 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdj9z\" (UniqueName: \"kubernetes.io/projected/6ca2b580-f2b5-4f61-9aa8-63b114198bb0-kube-api-access-bdj9z\") pod \"keystone-db-sync-9znjc\" (UID: \"6ca2b580-f2b5-4f61-9aa8-63b114198bb0\") " pod="openstack/keystone-db-sync-9znjc" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.184344 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnjj8\" (UniqueName: \"kubernetes.io/projected/beb0dafe-f425-4969-b586-29ca36b936f0-kube-api-access-gnjj8\") pod \"neutron-db-create-wsqj9\" (UID: \"beb0dafe-f425-4969-b586-29ca36b936f0\") " pod="openstack/neutron-db-create-wsqj9" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.205433 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnjj8\" (UniqueName: \"kubernetes.io/projected/beb0dafe-f425-4969-b586-29ca36b936f0-kube-api-access-gnjj8\") pod \"neutron-db-create-wsqj9\" (UID: \"beb0dafe-f425-4969-b586-29ca36b936f0\") " pod="openstack/neutron-db-create-wsqj9" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.286908 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ca2b580-f2b5-4f61-9aa8-63b114198bb0-config-data\") pod \"keystone-db-sync-9znjc\" (UID: \"6ca2b580-f2b5-4f61-9aa8-63b114198bb0\") " pod="openstack/keystone-db-sync-9znjc" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.287243 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdj9z\" (UniqueName: \"kubernetes.io/projected/6ca2b580-f2b5-4f61-9aa8-63b114198bb0-kube-api-access-bdj9z\") pod \"keystone-db-sync-9znjc\" (UID: \"6ca2b580-f2b5-4f61-9aa8-63b114198bb0\") " pod="openstack/keystone-db-sync-9znjc" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.287334 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca2b580-f2b5-4f61-9aa8-63b114198bb0-combined-ca-bundle\") pod \"keystone-db-sync-9znjc\" (UID: \"6ca2b580-f2b5-4f61-9aa8-63b114198bb0\") " pod="openstack/keystone-db-sync-9znjc" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.307107 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ca2b580-f2b5-4f61-9aa8-63b114198bb0-config-data\") pod \"keystone-db-sync-9znjc\" (UID: \"6ca2b580-f2b5-4f61-9aa8-63b114198bb0\") " pod="openstack/keystone-db-sync-9znjc" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.340257 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdj9z\" (UniqueName: \"kubernetes.io/projected/6ca2b580-f2b5-4f61-9aa8-63b114198bb0-kube-api-access-bdj9z\") pod \"keystone-db-sync-9znjc\" (UID: \"6ca2b580-f2b5-4f61-9aa8-63b114198bb0\") " pod="openstack/keystone-db-sync-9znjc" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.340412 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca2b580-f2b5-4f61-9aa8-63b114198bb0-combined-ca-bundle\") pod \"keystone-db-sync-9znjc\" (UID: \"6ca2b580-f2b5-4f61-9aa8-63b114198bb0\") " pod="openstack/keystone-db-sync-9znjc" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.344921 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-wsqj9" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.361041 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"aa1d7b22-8dac-42de-8241-b633bbe90691","Type":"ContainerStarted","Data":"65f259817feb0dbaabd9c6fa57300b99d9b07b82132825f3c60cf33eca6cfb5a"} Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.366358 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-x6xhq" event={"ID":"7542b422-3e5f-479c-ae79-d3209197d088","Type":"ContainerStarted","Data":"cff3994ec6b012ecee7843e150362954874e37adb08d33b5ce4811c1968777f0"} Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.415894 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=28.369643002 podStartE2EDuration="35.415874721s" podCreationTimestamp="2025-10-03 13:11:16 +0000 UTC" firstStartedPulling="2025-10-03 13:11:34.988485219 +0000 UTC m=+1230.786957403" lastFinishedPulling="2025-10-03 13:11:42.034716938 +0000 UTC m=+1237.833189122" observedRunningTime="2025-10-03 13:11:51.396163548 +0000 UTC m=+1247.194635732" watchObservedRunningTime="2025-10-03 13:11:51.415874721 +0000 UTC m=+1247.214346905" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.438881 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-9znjc" Oct 03 13:11:51 crc kubenswrapper[4578]: I1003 13:11:51.439459 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-x6xhq" podStartSLOduration=3.360476732 podStartE2EDuration="17.439426743s" podCreationTimestamp="2025-10-03 13:11:34 +0000 UTC" firstStartedPulling="2025-10-03 13:11:35.633218734 +0000 UTC m=+1231.431690918" lastFinishedPulling="2025-10-03 13:11:49.712168745 +0000 UTC m=+1245.510640929" observedRunningTime="2025-10-03 13:11:51.424019598 +0000 UTC m=+1247.222491782" watchObservedRunningTime="2025-10-03 13:11:51.439426743 +0000 UTC m=+1247.237898927" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:51.557429 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-twh27"] Oct 03 13:11:52 crc kubenswrapper[4578]: W1003 13:11:51.595819 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8999650e_96de_4898_9589_156a996042ed.slice/crio-c52ffa77e894c20785c700f8bcfc166b73c84a4a0cfc1f77404b44cb0a867d7b WatchSource:0}: Error finding container c52ffa77e894c20785c700f8bcfc166b73c84a4a0cfc1f77404b44cb0a867d7b: Status 404 returned error can't find the container with id c52ffa77e894c20785c700f8bcfc166b73c84a4a0cfc1f77404b44cb0a867d7b Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:51.772921 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-vzh7p"] Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:51.794814 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-s7mxj"] Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:51.796577 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:51.801062 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 03 13:11:52 crc kubenswrapper[4578]: W1003 13:11:51.807770 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1953665a_f20b_4620_8061_bc3e4a33aacd.slice/crio-cabee05b2c3ae5c47c2dbbca56be5a5b65e5867de13e211e0c5539dcb636c141 WatchSource:0}: Error finding container cabee05b2c3ae5c47c2dbbca56be5a5b65e5867de13e211e0c5539dcb636c141: Status 404 returned error can't find the container with id cabee05b2c3ae5c47c2dbbca56be5a5b65e5867de13e211e0c5539dcb636c141 Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:51.852614 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-s7mxj"] Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:51.906013 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-s7mxj\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:51.906098 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-config\") pod \"dnsmasq-dns-77585f5f8c-s7mxj\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:51.906122 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgl8s\" (UniqueName: \"kubernetes.io/projected/92cd4ef0-355f-4564-9590-6593431b678a-kube-api-access-tgl8s\") pod \"dnsmasq-dns-77585f5f8c-s7mxj\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:51.906230 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-s7mxj\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:51.906305 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-s7mxj\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:51.906328 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-s7mxj\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.007619 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-s7mxj\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.007961 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-s7mxj\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.007992 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-s7mxj\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.008050 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-config\") pod \"dnsmasq-dns-77585f5f8c-s7mxj\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.008074 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgl8s\" (UniqueName: \"kubernetes.io/projected/92cd4ef0-355f-4564-9590-6593431b678a-kube-api-access-tgl8s\") pod \"dnsmasq-dns-77585f5f8c-s7mxj\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.008143 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-s7mxj\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.008860 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-s7mxj\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.009454 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-config\") pod \"dnsmasq-dns-77585f5f8c-s7mxj\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.010079 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-s7mxj\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.010460 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-s7mxj\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.009621 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-s7mxj\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.037070 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgl8s\" (UniqueName: \"kubernetes.io/projected/92cd4ef0-355f-4564-9590-6593431b678a-kube-api-access-tgl8s\") pod \"dnsmasq-dns-77585f5f8c-s7mxj\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.179426 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.413490 4578 generic.go:334] "Generic (PLEG): container finished" podID="8999650e-96de-4898-9589-156a996042ed" containerID="5e1761875604e3e19eafa1974e4b52f4c326151f1a6c0a086240eecae5b56378" exitCode=0 Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.413572 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-twh27" event={"ID":"8999650e-96de-4898-9589-156a996042ed","Type":"ContainerDied","Data":"5e1761875604e3e19eafa1974e4b52f4c326151f1a6c0a086240eecae5b56378"} Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.416156 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-twh27" event={"ID":"8999650e-96de-4898-9589-156a996042ed","Type":"ContainerStarted","Data":"c52ffa77e894c20785c700f8bcfc166b73c84a4a0cfc1f77404b44cb0a867d7b"} Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.423218 4578 generic.go:334] "Generic (PLEG): container finished" podID="1953665a-f20b-4620-8061-bc3e4a33aacd" containerID="42be6a1ce882316a4c4f29463fafa8e22e0b267c6d13f8a2c788e0532bf6e4ad" exitCode=0 Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.424521 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-vzh7p" event={"ID":"1953665a-f20b-4620-8061-bc3e4a33aacd","Type":"ContainerDied","Data":"42be6a1ce882316a4c4f29463fafa8e22e0b267c6d13f8a2c788e0532bf6e4ad"} Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.424544 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-vzh7p" event={"ID":"1953665a-f20b-4620-8061-bc3e4a33aacd","Type":"ContainerStarted","Data":"cabee05b2c3ae5c47c2dbbca56be5a5b65e5867de13e211e0c5539dcb636c141"} Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.602280 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-9znjc"] Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.622165 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-wsqj9"] Oct 03 13:11:52 crc kubenswrapper[4578]: I1003 13:11:52.764545 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-s7mxj"] Oct 03 13:11:53 crc kubenswrapper[4578]: I1003 13:11:53.432992 4578 generic.go:334] "Generic (PLEG): container finished" podID="beb0dafe-f425-4969-b586-29ca36b936f0" containerID="7aa6fd6921703d6f49aedc22a01890766fc26f88404e28afecee61be9dca4e38" exitCode=0 Oct 03 13:11:53 crc kubenswrapper[4578]: I1003 13:11:53.433043 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-wsqj9" event={"ID":"beb0dafe-f425-4969-b586-29ca36b936f0","Type":"ContainerDied","Data":"7aa6fd6921703d6f49aedc22a01890766fc26f88404e28afecee61be9dca4e38"} Oct 03 13:11:53 crc kubenswrapper[4578]: I1003 13:11:53.433357 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-wsqj9" event={"ID":"beb0dafe-f425-4969-b586-29ca36b936f0","Type":"ContainerStarted","Data":"9700f57fee46304290c8b08bc1e87767581edb7ca42c3c30bef3d47f951c1681"} Oct 03 13:11:53 crc kubenswrapper[4578]: I1003 13:11:53.436326 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-9znjc" event={"ID":"6ca2b580-f2b5-4f61-9aa8-63b114198bb0","Type":"ContainerStarted","Data":"9484f5a436b28d339380a320a6e60c0a200216853e8be767443274f6dc3feb5a"} Oct 03 13:11:53 crc kubenswrapper[4578]: I1003 13:11:53.438217 4578 generic.go:334] "Generic (PLEG): container finished" podID="92cd4ef0-355f-4564-9590-6593431b678a" containerID="67712ace0dca9eebde84c2e8eef14d529293e75e8b113920ab2c4f941ca4c5d5" exitCode=0 Oct 03 13:11:53 crc kubenswrapper[4578]: I1003 13:11:53.438303 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" event={"ID":"92cd4ef0-355f-4564-9590-6593431b678a","Type":"ContainerDied","Data":"67712ace0dca9eebde84c2e8eef14d529293e75e8b113920ab2c4f941ca4c5d5"} Oct 03 13:11:53 crc kubenswrapper[4578]: I1003 13:11:53.438401 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" event={"ID":"92cd4ef0-355f-4564-9590-6593431b678a","Type":"ContainerStarted","Data":"8ae285be63c4d9907281b7363ed1fd92f0207006be3f49d57f867a9f236c9400"} Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.009012 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-vzh7p" Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.022085 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-twh27" Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.050527 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcrzl\" (UniqueName: \"kubernetes.io/projected/1953665a-f20b-4620-8061-bc3e4a33aacd-kube-api-access-xcrzl\") pod \"1953665a-f20b-4620-8061-bc3e4a33aacd\" (UID: \"1953665a-f20b-4620-8061-bc3e4a33aacd\") " Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.050698 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhqzm\" (UniqueName: \"kubernetes.io/projected/8999650e-96de-4898-9589-156a996042ed-kube-api-access-qhqzm\") pod \"8999650e-96de-4898-9589-156a996042ed\" (UID: \"8999650e-96de-4898-9589-156a996042ed\") " Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.070868 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1953665a-f20b-4620-8061-bc3e4a33aacd-kube-api-access-xcrzl" (OuterVolumeSpecName: "kube-api-access-xcrzl") pod "1953665a-f20b-4620-8061-bc3e4a33aacd" (UID: "1953665a-f20b-4620-8061-bc3e4a33aacd"). InnerVolumeSpecName "kube-api-access-xcrzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.070995 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8999650e-96de-4898-9589-156a996042ed-kube-api-access-qhqzm" (OuterVolumeSpecName: "kube-api-access-qhqzm") pod "8999650e-96de-4898-9589-156a996042ed" (UID: "8999650e-96de-4898-9589-156a996042ed"). InnerVolumeSpecName "kube-api-access-qhqzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.153015 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcrzl\" (UniqueName: \"kubernetes.io/projected/1953665a-f20b-4620-8061-bc3e4a33aacd-kube-api-access-xcrzl\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.153059 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhqzm\" (UniqueName: \"kubernetes.io/projected/8999650e-96de-4898-9589-156a996042ed-kube-api-access-qhqzm\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.454979 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" event={"ID":"92cd4ef0-355f-4564-9590-6593431b678a","Type":"ContainerStarted","Data":"67a6dda3e987feb1381d7012ff7fd320626f3627c9b84978669afe0518d9e937"} Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.455905 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.458997 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-twh27" event={"ID":"8999650e-96de-4898-9589-156a996042ed","Type":"ContainerDied","Data":"c52ffa77e894c20785c700f8bcfc166b73c84a4a0cfc1f77404b44cb0a867d7b"} Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.459024 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c52ffa77e894c20785c700f8bcfc166b73c84a4a0cfc1f77404b44cb0a867d7b" Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.459044 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-twh27" Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.460894 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-vzh7p" Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.461516 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-vzh7p" event={"ID":"1953665a-f20b-4620-8061-bc3e4a33aacd","Type":"ContainerDied","Data":"cabee05b2c3ae5c47c2dbbca56be5a5b65e5867de13e211e0c5539dcb636c141"} Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.461558 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cabee05b2c3ae5c47c2dbbca56be5a5b65e5867de13e211e0c5539dcb636c141" Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.482297 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" podStartSLOduration=3.482280125 podStartE2EDuration="3.482280125s" podCreationTimestamp="2025-10-03 13:11:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:11:54.473475177 +0000 UTC m=+1250.271947371" watchObservedRunningTime="2025-10-03 13:11:54.482280125 +0000 UTC m=+1250.280752309" Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.773493 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-wsqj9" Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.866551 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnjj8\" (UniqueName: \"kubernetes.io/projected/beb0dafe-f425-4969-b586-29ca36b936f0-kube-api-access-gnjj8\") pod \"beb0dafe-f425-4969-b586-29ca36b936f0\" (UID: \"beb0dafe-f425-4969-b586-29ca36b936f0\") " Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.872836 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beb0dafe-f425-4969-b586-29ca36b936f0-kube-api-access-gnjj8" (OuterVolumeSpecName: "kube-api-access-gnjj8") pod "beb0dafe-f425-4969-b586-29ca36b936f0" (UID: "beb0dafe-f425-4969-b586-29ca36b936f0"). InnerVolumeSpecName "kube-api-access-gnjj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:11:54 crc kubenswrapper[4578]: I1003 13:11:54.968886 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnjj8\" (UniqueName: \"kubernetes.io/projected/beb0dafe-f425-4969-b586-29ca36b936f0-kube-api-access-gnjj8\") on node \"crc\" DevicePath \"\"" Oct 03 13:11:55 crc kubenswrapper[4578]: I1003 13:11:55.472914 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-wsqj9" event={"ID":"beb0dafe-f425-4969-b586-29ca36b936f0","Type":"ContainerDied","Data":"9700f57fee46304290c8b08bc1e87767581edb7ca42c3c30bef3d47f951c1681"} Oct 03 13:11:55 crc kubenswrapper[4578]: I1003 13:11:55.472965 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9700f57fee46304290c8b08bc1e87767581edb7ca42c3c30bef3d47f951c1681" Oct 03 13:11:55 crc kubenswrapper[4578]: I1003 13:11:55.472945 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-wsqj9" Oct 03 13:11:59 crc kubenswrapper[4578]: I1003 13:11:59.511954 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-9znjc" event={"ID":"6ca2b580-f2b5-4f61-9aa8-63b114198bb0","Type":"ContainerStarted","Data":"97ba9510bcb1157d8687ff99064ab2bdcd89cf9a4d79983e75d04ef236835412"} Oct 03 13:11:59 crc kubenswrapper[4578]: I1003 13:11:59.526208 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-9znjc" podStartSLOduration=2.046088808 podStartE2EDuration="8.526192191s" podCreationTimestamp="2025-10-03 13:11:51 +0000 UTC" firstStartedPulling="2025-10-03 13:11:52.605973296 +0000 UTC m=+1248.404445480" lastFinishedPulling="2025-10-03 13:11:59.086076679 +0000 UTC m=+1254.884548863" observedRunningTime="2025-10-03 13:11:59.523823326 +0000 UTC m=+1255.322295510" watchObservedRunningTime="2025-10-03 13:11:59.526192191 +0000 UTC m=+1255.324664375" Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.520814 4578 generic.go:334] "Generic (PLEG): container finished" podID="7542b422-3e5f-479c-ae79-d3209197d088" containerID="cff3994ec6b012ecee7843e150362954874e37adb08d33b5ce4811c1968777f0" exitCode=0 Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.520909 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-x6xhq" event={"ID":"7542b422-3e5f-479c-ae79-d3209197d088","Type":"ContainerDied","Data":"cff3994ec6b012ecee7843e150362954874e37adb08d33b5ce4811c1968777f0"} Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.596548 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-2bce-account-create-zg8r5"] Oct 03 13:12:00 crc kubenswrapper[4578]: E1003 13:12:00.596970 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beb0dafe-f425-4969-b586-29ca36b936f0" containerName="mariadb-database-create" Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.596986 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="beb0dafe-f425-4969-b586-29ca36b936f0" containerName="mariadb-database-create" Oct 03 13:12:00 crc kubenswrapper[4578]: E1003 13:12:00.597011 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8999650e-96de-4898-9589-156a996042ed" containerName="mariadb-database-create" Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.597018 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="8999650e-96de-4898-9589-156a996042ed" containerName="mariadb-database-create" Oct 03 13:12:00 crc kubenswrapper[4578]: E1003 13:12:00.597029 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1953665a-f20b-4620-8061-bc3e4a33aacd" containerName="mariadb-database-create" Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.597037 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="1953665a-f20b-4620-8061-bc3e4a33aacd" containerName="mariadb-database-create" Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.597235 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="8999650e-96de-4898-9589-156a996042ed" containerName="mariadb-database-create" Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.597257 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="beb0dafe-f425-4969-b586-29ca36b936f0" containerName="mariadb-database-create" Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.597281 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="1953665a-f20b-4620-8061-bc3e4a33aacd" containerName="mariadb-database-create" Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.597964 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2bce-account-create-zg8r5" Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.602870 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.605385 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-2bce-account-create-zg8r5"] Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.690174 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hgrn\" (UniqueName: \"kubernetes.io/projected/b2537234-31d7-4970-a722-3e41b628008d-kube-api-access-6hgrn\") pod \"cinder-2bce-account-create-zg8r5\" (UID: \"b2537234-31d7-4970-a722-3e41b628008d\") " pod="openstack/cinder-2bce-account-create-zg8r5" Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.769468 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-f504-account-create-vrmwv"] Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.770482 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f504-account-create-vrmwv" Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.777425 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-f504-account-create-vrmwv"] Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.781467 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.792236 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hgrn\" (UniqueName: \"kubernetes.io/projected/b2537234-31d7-4970-a722-3e41b628008d-kube-api-access-6hgrn\") pod \"cinder-2bce-account-create-zg8r5\" (UID: \"b2537234-31d7-4970-a722-3e41b628008d\") " pod="openstack/cinder-2bce-account-create-zg8r5" Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.822020 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hgrn\" (UniqueName: \"kubernetes.io/projected/b2537234-31d7-4970-a722-3e41b628008d-kube-api-access-6hgrn\") pod \"cinder-2bce-account-create-zg8r5\" (UID: \"b2537234-31d7-4970-a722-3e41b628008d\") " pod="openstack/cinder-2bce-account-create-zg8r5" Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.893906 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwx6m\" (UniqueName: \"kubernetes.io/projected/3beaf630-6678-4fd2-8dc7-a6e50f51066c-kube-api-access-lwx6m\") pod \"barbican-f504-account-create-vrmwv\" (UID: \"3beaf630-6678-4fd2-8dc7-a6e50f51066c\") " pod="openstack/barbican-f504-account-create-vrmwv" Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.922051 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2bce-account-create-zg8r5" Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.996274 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwx6m\" (UniqueName: \"kubernetes.io/projected/3beaf630-6678-4fd2-8dc7-a6e50f51066c-kube-api-access-lwx6m\") pod \"barbican-f504-account-create-vrmwv\" (UID: \"3beaf630-6678-4fd2-8dc7-a6e50f51066c\") " pod="openstack/barbican-f504-account-create-vrmwv" Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.997031 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-ad4f-account-create-bp75p"] Oct 03 13:12:00 crc kubenswrapper[4578]: I1003 13:12:00.998686 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ad4f-account-create-bp75p" Oct 03 13:12:01 crc kubenswrapper[4578]: I1003 13:12:01.003025 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 03 13:12:01 crc kubenswrapper[4578]: I1003 13:12:01.011774 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ad4f-account-create-bp75p"] Oct 03 13:12:01 crc kubenswrapper[4578]: I1003 13:12:01.019833 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwx6m\" (UniqueName: \"kubernetes.io/projected/3beaf630-6678-4fd2-8dc7-a6e50f51066c-kube-api-access-lwx6m\") pod \"barbican-f504-account-create-vrmwv\" (UID: \"3beaf630-6678-4fd2-8dc7-a6e50f51066c\") " pod="openstack/barbican-f504-account-create-vrmwv" Oct 03 13:12:01 crc kubenswrapper[4578]: I1003 13:12:01.084358 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f504-account-create-vrmwv" Oct 03 13:12:01 crc kubenswrapper[4578]: I1003 13:12:01.097948 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pq4t\" (UniqueName: \"kubernetes.io/projected/b6bc24ff-faa7-41fa-a16c-ddd9533e39f4-kube-api-access-2pq4t\") pod \"neutron-ad4f-account-create-bp75p\" (UID: \"b6bc24ff-faa7-41fa-a16c-ddd9533e39f4\") " pod="openstack/neutron-ad4f-account-create-bp75p" Oct 03 13:12:01 crc kubenswrapper[4578]: I1003 13:12:01.199056 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pq4t\" (UniqueName: \"kubernetes.io/projected/b6bc24ff-faa7-41fa-a16c-ddd9533e39f4-kube-api-access-2pq4t\") pod \"neutron-ad4f-account-create-bp75p\" (UID: \"b6bc24ff-faa7-41fa-a16c-ddd9533e39f4\") " pod="openstack/neutron-ad4f-account-create-bp75p" Oct 03 13:12:01 crc kubenswrapper[4578]: I1003 13:12:01.228217 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pq4t\" (UniqueName: \"kubernetes.io/projected/b6bc24ff-faa7-41fa-a16c-ddd9533e39f4-kube-api-access-2pq4t\") pod \"neutron-ad4f-account-create-bp75p\" (UID: \"b6bc24ff-faa7-41fa-a16c-ddd9533e39f4\") " pod="openstack/neutron-ad4f-account-create-bp75p" Oct 03 13:12:01 crc kubenswrapper[4578]: I1003 13:12:01.321859 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ad4f-account-create-bp75p" Oct 03 13:12:01 crc kubenswrapper[4578]: I1003 13:12:01.433459 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-2bce-account-create-zg8r5"] Oct 03 13:12:01 crc kubenswrapper[4578]: W1003 13:12:01.440478 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2537234_31d7_4970_a722_3e41b628008d.slice/crio-441a23e0b5593feb7dbe27d15a2144f22f9e2a875b1436b76229feee665076a9 WatchSource:0}: Error finding container 441a23e0b5593feb7dbe27d15a2144f22f9e2a875b1436b76229feee665076a9: Status 404 returned error can't find the container with id 441a23e0b5593feb7dbe27d15a2144f22f9e2a875b1436b76229feee665076a9 Oct 03 13:12:01 crc kubenswrapper[4578]: I1003 13:12:01.541115 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2bce-account-create-zg8r5" event={"ID":"b2537234-31d7-4970-a722-3e41b628008d","Type":"ContainerStarted","Data":"441a23e0b5593feb7dbe27d15a2144f22f9e2a875b1436b76229feee665076a9"} Oct 03 13:12:01 crc kubenswrapper[4578]: I1003 13:12:01.550739 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-f504-account-create-vrmwv"] Oct 03 13:12:01 crc kubenswrapper[4578]: W1003 13:12:01.596149 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3beaf630_6678_4fd2_8dc7_a6e50f51066c.slice/crio-8543c36a60bf105b756bd75c212a9d93aacf9c4ec3e743ef294292b3a4109432 WatchSource:0}: Error finding container 8543c36a60bf105b756bd75c212a9d93aacf9c4ec3e743ef294292b3a4109432: Status 404 returned error can't find the container with id 8543c36a60bf105b756bd75c212a9d93aacf9c4ec3e743ef294292b3a4109432 Oct 03 13:12:01 crc kubenswrapper[4578]: I1003 13:12:01.783450 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ad4f-account-create-bp75p"] Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.093528 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-x6xhq" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.109489 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nl84s\" (UniqueName: \"kubernetes.io/projected/7542b422-3e5f-479c-ae79-d3209197d088-kube-api-access-nl84s\") pod \"7542b422-3e5f-479c-ae79-d3209197d088\" (UID: \"7542b422-3e5f-479c-ae79-d3209197d088\") " Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.109587 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7542b422-3e5f-479c-ae79-d3209197d088-db-sync-config-data\") pod \"7542b422-3e5f-479c-ae79-d3209197d088\" (UID: \"7542b422-3e5f-479c-ae79-d3209197d088\") " Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.109652 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7542b422-3e5f-479c-ae79-d3209197d088-config-data\") pod \"7542b422-3e5f-479c-ae79-d3209197d088\" (UID: \"7542b422-3e5f-479c-ae79-d3209197d088\") " Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.168297 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7542b422-3e5f-479c-ae79-d3209197d088-kube-api-access-nl84s" (OuterVolumeSpecName: "kube-api-access-nl84s") pod "7542b422-3e5f-479c-ae79-d3209197d088" (UID: "7542b422-3e5f-479c-ae79-d3209197d088"). InnerVolumeSpecName "kube-api-access-nl84s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.173519 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7542b422-3e5f-479c-ae79-d3209197d088-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7542b422-3e5f-479c-ae79-d3209197d088" (UID: "7542b422-3e5f-479c-ae79-d3209197d088"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.180820 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.200914 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7542b422-3e5f-479c-ae79-d3209197d088-config-data" (OuterVolumeSpecName: "config-data") pod "7542b422-3e5f-479c-ae79-d3209197d088" (UID: "7542b422-3e5f-479c-ae79-d3209197d088"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.211339 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7542b422-3e5f-479c-ae79-d3209197d088-combined-ca-bundle\") pod \"7542b422-3e5f-479c-ae79-d3209197d088\" (UID: \"7542b422-3e5f-479c-ae79-d3209197d088\") " Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.211943 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nl84s\" (UniqueName: \"kubernetes.io/projected/7542b422-3e5f-479c-ae79-d3209197d088-kube-api-access-nl84s\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.211965 4578 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7542b422-3e5f-479c-ae79-d3209197d088-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.211977 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7542b422-3e5f-479c-ae79-d3209197d088-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.252065 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-k9bht"] Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.253008 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-k9bht" podUID="68090354-cfd7-472a-bca7-a9b06f40957f" containerName="dnsmasq-dns" containerID="cri-o://2ce8a6bad39cb722b50bc5ba9cecd313e4b8e85b77f30dfbe0796b8df15b1ffb" gracePeriod=10 Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.296216 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7542b422-3e5f-479c-ae79-d3209197d088-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7542b422-3e5f-479c-ae79-d3209197d088" (UID: "7542b422-3e5f-479c-ae79-d3209197d088"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.313656 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7542b422-3e5f-479c-ae79-d3209197d088-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.556233 4578 generic.go:334] "Generic (PLEG): container finished" podID="b2537234-31d7-4970-a722-3e41b628008d" containerID="86369178d83c39ee274194d59242ab148400430763c8e37c53b0b3a25eef58a8" exitCode=0 Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.556334 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2bce-account-create-zg8r5" event={"ID":"b2537234-31d7-4970-a722-3e41b628008d","Type":"ContainerDied","Data":"86369178d83c39ee274194d59242ab148400430763c8e37c53b0b3a25eef58a8"} Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.562097 4578 generic.go:334] "Generic (PLEG): container finished" podID="68090354-cfd7-472a-bca7-a9b06f40957f" containerID="2ce8a6bad39cb722b50bc5ba9cecd313e4b8e85b77f30dfbe0796b8df15b1ffb" exitCode=0 Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.562187 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-k9bht" event={"ID":"68090354-cfd7-472a-bca7-a9b06f40957f","Type":"ContainerDied","Data":"2ce8a6bad39cb722b50bc5ba9cecd313e4b8e85b77f30dfbe0796b8df15b1ffb"} Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.565738 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-x6xhq" event={"ID":"7542b422-3e5f-479c-ae79-d3209197d088","Type":"ContainerDied","Data":"2734b6120170924d6cdd842bba117d13d842570c3725ce053a5c8bc85fc09f19"} Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.565773 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2734b6120170924d6cdd842bba117d13d842570c3725ce053a5c8bc85fc09f19" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.565838 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-x6xhq" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.569947 4578 generic.go:334] "Generic (PLEG): container finished" podID="3beaf630-6678-4fd2-8dc7-a6e50f51066c" containerID="50e29480e63e2b503ccb4c85c9df0b0ac393219bfa25c4c5027a0e465c8efabd" exitCode=0 Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.569990 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-f504-account-create-vrmwv" event={"ID":"3beaf630-6678-4fd2-8dc7-a6e50f51066c","Type":"ContainerDied","Data":"50e29480e63e2b503ccb4c85c9df0b0ac393219bfa25c4c5027a0e465c8efabd"} Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.570029 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-f504-account-create-vrmwv" event={"ID":"3beaf630-6678-4fd2-8dc7-a6e50f51066c","Type":"ContainerStarted","Data":"8543c36a60bf105b756bd75c212a9d93aacf9c4ec3e743ef294292b3a4109432"} Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.579402 4578 generic.go:334] "Generic (PLEG): container finished" podID="b6bc24ff-faa7-41fa-a16c-ddd9533e39f4" containerID="5f0c1811e453f528295066dc5922a14afc246f86e36a0320719f0eac013a3cc8" exitCode=0 Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.579476 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ad4f-account-create-bp75p" event={"ID":"b6bc24ff-faa7-41fa-a16c-ddd9533e39f4","Type":"ContainerDied","Data":"5f0c1811e453f528295066dc5922a14afc246f86e36a0320719f0eac013a3cc8"} Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.579507 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ad4f-account-create-bp75p" event={"ID":"b6bc24ff-faa7-41fa-a16c-ddd9533e39f4","Type":"ContainerStarted","Data":"326ca8a02dddd6d5136ec0334e9b9054a1305730608df3f89a5bf14536b4e6f7"} Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.624546 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.719229 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-ovsdbserver-sb\") pod \"68090354-cfd7-472a-bca7-a9b06f40957f\" (UID: \"68090354-cfd7-472a-bca7-a9b06f40957f\") " Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.719271 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-dns-svc\") pod \"68090354-cfd7-472a-bca7-a9b06f40957f\" (UID: \"68090354-cfd7-472a-bca7-a9b06f40957f\") " Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.719398 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-ovsdbserver-nb\") pod \"68090354-cfd7-472a-bca7-a9b06f40957f\" (UID: \"68090354-cfd7-472a-bca7-a9b06f40957f\") " Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.719446 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4l8k\" (UniqueName: \"kubernetes.io/projected/68090354-cfd7-472a-bca7-a9b06f40957f-kube-api-access-v4l8k\") pod \"68090354-cfd7-472a-bca7-a9b06f40957f\" (UID: \"68090354-cfd7-472a-bca7-a9b06f40957f\") " Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.719467 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-config\") pod \"68090354-cfd7-472a-bca7-a9b06f40957f\" (UID: \"68090354-cfd7-472a-bca7-a9b06f40957f\") " Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.730322 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68090354-cfd7-472a-bca7-a9b06f40957f-kube-api-access-v4l8k" (OuterVolumeSpecName: "kube-api-access-v4l8k") pod "68090354-cfd7-472a-bca7-a9b06f40957f" (UID: "68090354-cfd7-472a-bca7-a9b06f40957f"). InnerVolumeSpecName "kube-api-access-v4l8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.761897 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "68090354-cfd7-472a-bca7-a9b06f40957f" (UID: "68090354-cfd7-472a-bca7-a9b06f40957f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.765202 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-config" (OuterVolumeSpecName: "config") pod "68090354-cfd7-472a-bca7-a9b06f40957f" (UID: "68090354-cfd7-472a-bca7-a9b06f40957f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.770706 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "68090354-cfd7-472a-bca7-a9b06f40957f" (UID: "68090354-cfd7-472a-bca7-a9b06f40957f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.788795 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "68090354-cfd7-472a-bca7-a9b06f40957f" (UID: "68090354-cfd7-472a-bca7-a9b06f40957f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.821534 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4l8k\" (UniqueName: \"kubernetes.io/projected/68090354-cfd7-472a-bca7-a9b06f40957f-kube-api-access-v4l8k\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.821569 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.821580 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.821587 4578 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.821597 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68090354-cfd7-472a-bca7-a9b06f40957f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.989476 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-8k8j7"] Oct 03 13:12:02 crc kubenswrapper[4578]: E1003 13:12:02.989928 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68090354-cfd7-472a-bca7-a9b06f40957f" containerName="init" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.989948 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="68090354-cfd7-472a-bca7-a9b06f40957f" containerName="init" Oct 03 13:12:02 crc kubenswrapper[4578]: E1003 13:12:02.989963 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7542b422-3e5f-479c-ae79-d3209197d088" containerName="glance-db-sync" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.989969 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="7542b422-3e5f-479c-ae79-d3209197d088" containerName="glance-db-sync" Oct 03 13:12:02 crc kubenswrapper[4578]: E1003 13:12:02.989990 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68090354-cfd7-472a-bca7-a9b06f40957f" containerName="dnsmasq-dns" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.989997 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="68090354-cfd7-472a-bca7-a9b06f40957f" containerName="dnsmasq-dns" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.990158 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="68090354-cfd7-472a-bca7-a9b06f40957f" containerName="dnsmasq-dns" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.990172 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="7542b422-3e5f-479c-ae79-d3209197d088" containerName="glance-db-sync" Oct 03 13:12:02 crc kubenswrapper[4578]: I1003 13:12:02.991218 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.002866 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-8k8j7"] Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.027228 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-8k8j7\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.027268 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-config\") pod \"dnsmasq-dns-7ff5475cc9-8k8j7\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.027321 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-8k8j7\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.027348 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-8k8j7\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.027372 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-8k8j7\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.027411 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn9pt\" (UniqueName: \"kubernetes.io/projected/5f1301ed-9d0f-4043-a967-2718ce351042-kube-api-access-vn9pt\") pod \"dnsmasq-dns-7ff5475cc9-8k8j7\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.128565 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn9pt\" (UniqueName: \"kubernetes.io/projected/5f1301ed-9d0f-4043-a967-2718ce351042-kube-api-access-vn9pt\") pod \"dnsmasq-dns-7ff5475cc9-8k8j7\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.128998 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-8k8j7\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.129019 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-config\") pod \"dnsmasq-dns-7ff5475cc9-8k8j7\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.129071 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-8k8j7\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.129094 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-8k8j7\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.129122 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-8k8j7\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.129794 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-config\") pod \"dnsmasq-dns-7ff5475cc9-8k8j7\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.129926 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-8k8j7\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.130349 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-8k8j7\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.130661 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-8k8j7\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.130815 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-8k8j7\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.146510 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn9pt\" (UniqueName: \"kubernetes.io/projected/5f1301ed-9d0f-4043-a967-2718ce351042-kube-api-access-vn9pt\") pod \"dnsmasq-dns-7ff5475cc9-8k8j7\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.322248 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.591571 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-k9bht" event={"ID":"68090354-cfd7-472a-bca7-a9b06f40957f","Type":"ContainerDied","Data":"48f8191f2b4fcb115efbc5d51eaa83d11d80093fbc65c5e774f8a97fddce4531"} Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.592070 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-k9bht" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.592705 4578 scope.go:117] "RemoveContainer" containerID="2ce8a6bad39cb722b50bc5ba9cecd313e4b8e85b77f30dfbe0796b8df15b1ffb" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.600963 4578 generic.go:334] "Generic (PLEG): container finished" podID="6ca2b580-f2b5-4f61-9aa8-63b114198bb0" containerID="97ba9510bcb1157d8687ff99064ab2bdcd89cf9a4d79983e75d04ef236835412" exitCode=0 Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.601050 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-9znjc" event={"ID":"6ca2b580-f2b5-4f61-9aa8-63b114198bb0","Type":"ContainerDied","Data":"97ba9510bcb1157d8687ff99064ab2bdcd89cf9a4d79983e75d04ef236835412"} Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.617612 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-k9bht"] Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.634417 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-k9bht"] Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.634622 4578 scope.go:117] "RemoveContainer" containerID="157d59c22da841c160afb1f51aa48c3d37e8f41c5a8c88abf633a4e7791a7eb1" Oct 03 13:12:03 crc kubenswrapper[4578]: I1003 13:12:03.879370 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-8k8j7"] Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.051112 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2bce-account-create-zg8r5" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.152661 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f504-account-create-vrmwv" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.158916 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hgrn\" (UniqueName: \"kubernetes.io/projected/b2537234-31d7-4970-a722-3e41b628008d-kube-api-access-6hgrn\") pod \"b2537234-31d7-4970-a722-3e41b628008d\" (UID: \"b2537234-31d7-4970-a722-3e41b628008d\") " Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.163097 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2537234-31d7-4970-a722-3e41b628008d-kube-api-access-6hgrn" (OuterVolumeSpecName: "kube-api-access-6hgrn") pod "b2537234-31d7-4970-a722-3e41b628008d" (UID: "b2537234-31d7-4970-a722-3e41b628008d"). InnerVolumeSpecName "kube-api-access-6hgrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.180097 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ad4f-account-create-bp75p" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.259995 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwx6m\" (UniqueName: \"kubernetes.io/projected/3beaf630-6678-4fd2-8dc7-a6e50f51066c-kube-api-access-lwx6m\") pod \"3beaf630-6678-4fd2-8dc7-a6e50f51066c\" (UID: \"3beaf630-6678-4fd2-8dc7-a6e50f51066c\") " Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.260144 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pq4t\" (UniqueName: \"kubernetes.io/projected/b6bc24ff-faa7-41fa-a16c-ddd9533e39f4-kube-api-access-2pq4t\") pod \"b6bc24ff-faa7-41fa-a16c-ddd9533e39f4\" (UID: \"b6bc24ff-faa7-41fa-a16c-ddd9533e39f4\") " Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.260667 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hgrn\" (UniqueName: \"kubernetes.io/projected/b2537234-31d7-4970-a722-3e41b628008d-kube-api-access-6hgrn\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.264303 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6bc24ff-faa7-41fa-a16c-ddd9533e39f4-kube-api-access-2pq4t" (OuterVolumeSpecName: "kube-api-access-2pq4t") pod "b6bc24ff-faa7-41fa-a16c-ddd9533e39f4" (UID: "b6bc24ff-faa7-41fa-a16c-ddd9533e39f4"). InnerVolumeSpecName "kube-api-access-2pq4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.265440 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3beaf630-6678-4fd2-8dc7-a6e50f51066c-kube-api-access-lwx6m" (OuterVolumeSpecName: "kube-api-access-lwx6m") pod "3beaf630-6678-4fd2-8dc7-a6e50f51066c" (UID: "3beaf630-6678-4fd2-8dc7-a6e50f51066c"). InnerVolumeSpecName "kube-api-access-lwx6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.362086 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pq4t\" (UniqueName: \"kubernetes.io/projected/b6bc24ff-faa7-41fa-a16c-ddd9533e39f4-kube-api-access-2pq4t\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.362122 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwx6m\" (UniqueName: \"kubernetes.io/projected/3beaf630-6678-4fd2-8dc7-a6e50f51066c-kube-api-access-lwx6m\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.608933 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" event={"ID":"5f1301ed-9d0f-4043-a967-2718ce351042","Type":"ContainerStarted","Data":"f50d837f103d8aae29adaed2d507f4d788c540cc67584a6f62e6ff1075ee03e5"} Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.611213 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f504-account-create-vrmwv" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.611204 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-f504-account-create-vrmwv" event={"ID":"3beaf630-6678-4fd2-8dc7-a6e50f51066c","Type":"ContainerDied","Data":"8543c36a60bf105b756bd75c212a9d93aacf9c4ec3e743ef294292b3a4109432"} Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.611333 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8543c36a60bf105b756bd75c212a9d93aacf9c4ec3e743ef294292b3a4109432" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.612786 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ad4f-account-create-bp75p" event={"ID":"b6bc24ff-faa7-41fa-a16c-ddd9533e39f4","Type":"ContainerDied","Data":"326ca8a02dddd6d5136ec0334e9b9054a1305730608df3f89a5bf14536b4e6f7"} Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.612809 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="326ca8a02dddd6d5136ec0334e9b9054a1305730608df3f89a5bf14536b4e6f7" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.612854 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ad4f-account-create-bp75p" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.615241 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-2bce-account-create-zg8r5" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.615241 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-2bce-account-create-zg8r5" event={"ID":"b2537234-31d7-4970-a722-3e41b628008d","Type":"ContainerDied","Data":"441a23e0b5593feb7dbe27d15a2144f22f9e2a875b1436b76229feee665076a9"} Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.615369 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="441a23e0b5593feb7dbe27d15a2144f22f9e2a875b1436b76229feee665076a9" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.840816 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-9znjc" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.872421 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdj9z\" (UniqueName: \"kubernetes.io/projected/6ca2b580-f2b5-4f61-9aa8-63b114198bb0-kube-api-access-bdj9z\") pod \"6ca2b580-f2b5-4f61-9aa8-63b114198bb0\" (UID: \"6ca2b580-f2b5-4f61-9aa8-63b114198bb0\") " Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.872644 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca2b580-f2b5-4f61-9aa8-63b114198bb0-combined-ca-bundle\") pod \"6ca2b580-f2b5-4f61-9aa8-63b114198bb0\" (UID: \"6ca2b580-f2b5-4f61-9aa8-63b114198bb0\") " Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.872827 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ca2b580-f2b5-4f61-9aa8-63b114198bb0-config-data\") pod \"6ca2b580-f2b5-4f61-9aa8-63b114198bb0\" (UID: \"6ca2b580-f2b5-4f61-9aa8-63b114198bb0\") " Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.887082 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ca2b580-f2b5-4f61-9aa8-63b114198bb0-kube-api-access-bdj9z" (OuterVolumeSpecName: "kube-api-access-bdj9z") pod "6ca2b580-f2b5-4f61-9aa8-63b114198bb0" (UID: "6ca2b580-f2b5-4f61-9aa8-63b114198bb0"). InnerVolumeSpecName "kube-api-access-bdj9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.931059 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ca2b580-f2b5-4f61-9aa8-63b114198bb0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ca2b580-f2b5-4f61-9aa8-63b114198bb0" (UID: "6ca2b580-f2b5-4f61-9aa8-63b114198bb0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.949656 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68090354-cfd7-472a-bca7-a9b06f40957f" path="/var/lib/kubelet/pods/68090354-cfd7-472a-bca7-a9b06f40957f/volumes" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.963024 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ca2b580-f2b5-4f61-9aa8-63b114198bb0-config-data" (OuterVolumeSpecName: "config-data") pod "6ca2b580-f2b5-4f61-9aa8-63b114198bb0" (UID: "6ca2b580-f2b5-4f61-9aa8-63b114198bb0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.981486 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdj9z\" (UniqueName: \"kubernetes.io/projected/6ca2b580-f2b5-4f61-9aa8-63b114198bb0-kube-api-access-bdj9z\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.981514 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca2b580-f2b5-4f61-9aa8-63b114198bb0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:04 crc kubenswrapper[4578]: I1003 13:12:04.981523 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ca2b580-f2b5-4f61-9aa8-63b114198bb0-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.091837 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.091898 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.628550 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-9znjc" event={"ID":"6ca2b580-f2b5-4f61-9aa8-63b114198bb0","Type":"ContainerDied","Data":"9484f5a436b28d339380a320a6e60c0a200216853e8be767443274f6dc3feb5a"} Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.628948 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9484f5a436b28d339380a320a6e60c0a200216853e8be767443274f6dc3feb5a" Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.628615 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-9znjc" Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.855457 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-8k8j7"] Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.891318 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q"] Oct 03 13:12:05 crc kubenswrapper[4578]: E1003 13:12:05.891646 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ca2b580-f2b5-4f61-9aa8-63b114198bb0" containerName="keystone-db-sync" Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.891662 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ca2b580-f2b5-4f61-9aa8-63b114198bb0" containerName="keystone-db-sync" Oct 03 13:12:05 crc kubenswrapper[4578]: E1003 13:12:05.891677 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2537234-31d7-4970-a722-3e41b628008d" containerName="mariadb-account-create" Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.891685 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2537234-31d7-4970-a722-3e41b628008d" containerName="mariadb-account-create" Oct 03 13:12:05 crc kubenswrapper[4578]: E1003 13:12:05.891705 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6bc24ff-faa7-41fa-a16c-ddd9533e39f4" containerName="mariadb-account-create" Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.891712 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6bc24ff-faa7-41fa-a16c-ddd9533e39f4" containerName="mariadb-account-create" Oct 03 13:12:05 crc kubenswrapper[4578]: E1003 13:12:05.891733 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3beaf630-6678-4fd2-8dc7-a6e50f51066c" containerName="mariadb-account-create" Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.891739 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="3beaf630-6678-4fd2-8dc7-a6e50f51066c" containerName="mariadb-account-create" Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.891887 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6bc24ff-faa7-41fa-a16c-ddd9533e39f4" containerName="mariadb-account-create" Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.891908 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2537234-31d7-4970-a722-3e41b628008d" containerName="mariadb-account-create" Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.891920 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ca2b580-f2b5-4f61-9aa8-63b114198bb0" containerName="keystone-db-sync" Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.891932 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="3beaf630-6678-4fd2-8dc7-a6e50f51066c" containerName="mariadb-account-create" Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.892790 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.909523 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q"] Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.958267 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-zmn6h"] Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.959249 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.961122 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.961385 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.961562 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.961815 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wxw7h" Oct 03 13:12:05 crc kubenswrapper[4578]: I1003 13:12:05.977441 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-zmn6h"] Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.000859 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-scripts\") pod \"keystone-bootstrap-zmn6h\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.000911 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-qlj9q\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.000936 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-combined-ca-bundle\") pod \"keystone-bootstrap-zmn6h\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.000963 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-config-data\") pod \"keystone-bootstrap-zmn6h\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.000988 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-qlj9q\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.005956 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-662gl\" (UniqueName: \"kubernetes.io/projected/3de77cab-fde4-438d-956c-8381b6d003c2-kube-api-access-662gl\") pod \"keystone-bootstrap-zmn6h\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.005998 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-config\") pod \"dnsmasq-dns-5c5cc7c5ff-qlj9q\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.006062 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-qlj9q\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.006083 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-fernet-keys\") pod \"keystone-bootstrap-zmn6h\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.006184 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-qlj9q\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.006243 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2p9s\" (UniqueName: \"kubernetes.io/projected/46649c43-396c-49fc-a0be-982d1805a483-kube-api-access-d2p9s\") pod \"dnsmasq-dns-5c5cc7c5ff-qlj9q\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.006264 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-credential-keys\") pod \"keystone-bootstrap-zmn6h\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.107951 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-config-data\") pod \"keystone-bootstrap-zmn6h\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.108000 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-qlj9q\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.108023 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-662gl\" (UniqueName: \"kubernetes.io/projected/3de77cab-fde4-438d-956c-8381b6d003c2-kube-api-access-662gl\") pod \"keystone-bootstrap-zmn6h\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.108047 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-config\") pod \"dnsmasq-dns-5c5cc7c5ff-qlj9q\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.108088 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-qlj9q\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.108107 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-fernet-keys\") pod \"keystone-bootstrap-zmn6h\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.108164 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-qlj9q\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.108212 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2p9s\" (UniqueName: \"kubernetes.io/projected/46649c43-396c-49fc-a0be-982d1805a483-kube-api-access-d2p9s\") pod \"dnsmasq-dns-5c5cc7c5ff-qlj9q\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.108237 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-credential-keys\") pod \"keystone-bootstrap-zmn6h\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.108263 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-scripts\") pod \"keystone-bootstrap-zmn6h\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.108278 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-qlj9q\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.108299 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-combined-ca-bundle\") pod \"keystone-bootstrap-zmn6h\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.108936 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-qlj9q\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.108951 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-qlj9q\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.109547 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-config\") pod \"dnsmasq-dns-5c5cc7c5ff-qlj9q\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.109865 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-qlj9q\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.110103 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-qlj9q\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.118237 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-fernet-keys\") pod \"keystone-bootstrap-zmn6h\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.118411 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-scripts\") pod \"keystone-bootstrap-zmn6h\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.118532 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-config-data\") pod \"keystone-bootstrap-zmn6h\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.124258 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-credential-keys\") pod \"keystone-bootstrap-zmn6h\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.131378 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-q6ppj"] Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.132415 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.136953 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-p5nqp" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.137139 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.137325 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.144510 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-662gl\" (UniqueName: \"kubernetes.io/projected/3de77cab-fde4-438d-956c-8381b6d003c2-kube-api-access-662gl\") pod \"keystone-bootstrap-zmn6h\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.152464 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-combined-ca-bundle\") pod \"keystone-bootstrap-zmn6h\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.157948 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2p9s\" (UniqueName: \"kubernetes.io/projected/46649c43-396c-49fc-a0be-982d1805a483-kube-api-access-d2p9s\") pod \"dnsmasq-dns-5c5cc7c5ff-qlj9q\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.181139 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7c684ff4d9-2h9cq"] Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.182517 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c684ff4d9-2h9cq" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.187002 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.188859 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-nchlm" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.189184 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.189438 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.211466 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-horizon-secret-key\") pod \"horizon-7c684ff4d9-2h9cq\" (UID: \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\") " pod="openstack/horizon-7c684ff4d9-2h9cq" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.211878 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-logs\") pod \"horizon-7c684ff4d9-2h9cq\" (UID: \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\") " pod="openstack/horizon-7c684ff4d9-2h9cq" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.211903 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-config-data\") pod \"cinder-db-sync-q6ppj\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.211923 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b310034-f02f-48ac-a47a-4f41f15379a5-etc-machine-id\") pod \"cinder-db-sync-q6ppj\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.211952 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55whk\" (UniqueName: \"kubernetes.io/projected/6b310034-f02f-48ac-a47a-4f41f15379a5-kube-api-access-55whk\") pod \"cinder-db-sync-q6ppj\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.211968 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-db-sync-config-data\") pod \"cinder-db-sync-q6ppj\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.211985 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-config-data\") pod \"horizon-7c684ff4d9-2h9cq\" (UID: \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\") " pod="openstack/horizon-7c684ff4d9-2h9cq" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.212005 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-combined-ca-bundle\") pod \"cinder-db-sync-q6ppj\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.212057 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-scripts\") pod \"cinder-db-sync-q6ppj\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.212098 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-scripts\") pod \"horizon-7c684ff4d9-2h9cq\" (UID: \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\") " pod="openstack/horizon-7c684ff4d9-2h9cq" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.212116 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfkx6\" (UniqueName: \"kubernetes.io/projected/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-kube-api-access-gfkx6\") pod \"horizon-7c684ff4d9-2h9cq\" (UID: \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\") " pod="openstack/horizon-7c684ff4d9-2h9cq" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.220755 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c684ff4d9-2h9cq"] Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.222551 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.247463 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-q6ppj"] Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.277690 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.313611 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-scripts\") pod \"horizon-7c684ff4d9-2h9cq\" (UID: \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\") " pod="openstack/horizon-7c684ff4d9-2h9cq" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.313671 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfkx6\" (UniqueName: \"kubernetes.io/projected/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-kube-api-access-gfkx6\") pod \"horizon-7c684ff4d9-2h9cq\" (UID: \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\") " pod="openstack/horizon-7c684ff4d9-2h9cq" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.313726 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-horizon-secret-key\") pod \"horizon-7c684ff4d9-2h9cq\" (UID: \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\") " pod="openstack/horizon-7c684ff4d9-2h9cq" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.313749 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-logs\") pod \"horizon-7c684ff4d9-2h9cq\" (UID: \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\") " pod="openstack/horizon-7c684ff4d9-2h9cq" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.313766 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-config-data\") pod \"cinder-db-sync-q6ppj\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.313782 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b310034-f02f-48ac-a47a-4f41f15379a5-etc-machine-id\") pod \"cinder-db-sync-q6ppj\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.313802 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55whk\" (UniqueName: \"kubernetes.io/projected/6b310034-f02f-48ac-a47a-4f41f15379a5-kube-api-access-55whk\") pod \"cinder-db-sync-q6ppj\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.313820 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-db-sync-config-data\") pod \"cinder-db-sync-q6ppj\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.313836 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-config-data\") pod \"horizon-7c684ff4d9-2h9cq\" (UID: \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\") " pod="openstack/horizon-7c684ff4d9-2h9cq" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.313856 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-combined-ca-bundle\") pod \"cinder-db-sync-q6ppj\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.313902 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-scripts\") pod \"cinder-db-sync-q6ppj\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.314727 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b310034-f02f-48ac-a47a-4f41f15379a5-etc-machine-id\") pod \"cinder-db-sync-q6ppj\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.315873 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-scripts\") pod \"horizon-7c684ff4d9-2h9cq\" (UID: \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\") " pod="openstack/horizon-7c684ff4d9-2h9cq" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.316231 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-config-data\") pod \"horizon-7c684ff4d9-2h9cq\" (UID: \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\") " pod="openstack/horizon-7c684ff4d9-2h9cq" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.317278 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-logs\") pod \"horizon-7c684ff4d9-2h9cq\" (UID: \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\") " pod="openstack/horizon-7c684ff4d9-2h9cq" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.325604 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-scripts\") pod \"cinder-db-sync-q6ppj\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.331362 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-db-sync-config-data\") pod \"cinder-db-sync-q6ppj\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.331449 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-combined-ca-bundle\") pod \"cinder-db-sync-q6ppj\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.338386 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-horizon-secret-key\") pod \"horizon-7c684ff4d9-2h9cq\" (UID: \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\") " pod="openstack/horizon-7c684ff4d9-2h9cq" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.339971 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-config-data\") pod \"cinder-db-sync-q6ppj\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.372787 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfkx6\" (UniqueName: \"kubernetes.io/projected/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-kube-api-access-gfkx6\") pod \"horizon-7c684ff4d9-2h9cq\" (UID: \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\") " pod="openstack/horizon-7c684ff4d9-2h9cq" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.381783 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.392847 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.397515 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55whk\" (UniqueName: \"kubernetes.io/projected/6b310034-f02f-48ac-a47a-4f41f15379a5-kube-api-access-55whk\") pod \"cinder-db-sync-q6ppj\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.404038 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.404160 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.414971 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6228c849-5824-4a21-8318-754c65d79a2e-log-httpd\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.415046 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-scripts\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.415069 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-config-data\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.415090 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbm4c\" (UniqueName: \"kubernetes.io/projected/6228c849-5824-4a21-8318-754c65d79a2e-kube-api-access-tbm4c\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.415127 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.415194 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.415238 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6228c849-5824-4a21-8318-754c65d79a2e-run-httpd\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.422260 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.515314 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.516614 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-scripts\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.516680 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-config-data\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.516716 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbm4c\" (UniqueName: \"kubernetes.io/projected/6228c849-5824-4a21-8318-754c65d79a2e-kube-api-access-tbm4c\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.516750 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.516790 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.516821 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6228c849-5824-4a21-8318-754c65d79a2e-run-httpd\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.516848 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6228c849-5824-4a21-8318-754c65d79a2e-log-httpd\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.517453 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6228c849-5824-4a21-8318-754c65d79a2e-run-httpd\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.517233 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6228c849-5824-4a21-8318-754c65d79a2e-log-httpd\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.521345 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-scripts\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.525799 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.535101 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-config-data\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.535515 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c684ff4d9-2h9cq" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.540125 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.592364 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbm4c\" (UniqueName: \"kubernetes.io/projected/6228c849-5824-4a21-8318-754c65d79a2e-kube-api-access-tbm4c\") pod \"ceilometer-0\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.628801 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-85psh"] Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.649665 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-85psh" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.684099 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.694499 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-wbbsn" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.712273 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.737822 4578 generic.go:334] "Generic (PLEG): container finished" podID="5f1301ed-9d0f-4043-a967-2718ce351042" containerID="d796bd865b5be852491ecd46807dc296d85ff5a178c8fb01953ee842127af9c5" exitCode=0 Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.737904 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" event={"ID":"5f1301ed-9d0f-4043-a967-2718ce351042","Type":"ContainerDied","Data":"d796bd865b5be852491ecd46807dc296d85ff5a178c8fb01953ee842127af9c5"} Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.738734 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.785820 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7fb966687c-gnkj8"] Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.787212 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fb966687c-gnkj8" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.816910 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-85psh"] Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.851610 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924-config\") pod \"neutron-db-sync-85psh\" (UID: \"0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924\") " pod="openstack/neutron-db-sync-85psh" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.851828 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924-combined-ca-bundle\") pod \"neutron-db-sync-85psh\" (UID: \"0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924\") " pod="openstack/neutron-db-sync-85psh" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.852015 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7zk9\" (UniqueName: \"kubernetes.io/projected/0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924-kube-api-access-d7zk9\") pod \"neutron-db-sync-85psh\" (UID: \"0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924\") " pod="openstack/neutron-db-sync-85psh" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.866320 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7fb966687c-gnkj8"] Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.899229 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-j9scz"] Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.928732 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-j9scz" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.944592 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8fxh8" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.945282 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.953343 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7zk9\" (UniqueName: \"kubernetes.io/projected/0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924-kube-api-access-d7zk9\") pod \"neutron-db-sync-85psh\" (UID: \"0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924\") " pod="openstack/neutron-db-sync-85psh" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.953410 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924-config\") pod \"neutron-db-sync-85psh\" (UID: \"0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924\") " pod="openstack/neutron-db-sync-85psh" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.953448 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-scripts\") pod \"horizon-7fb966687c-gnkj8\" (UID: \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\") " pod="openstack/horizon-7fb966687c-gnkj8" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.953471 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w85j6\" (UniqueName: \"kubernetes.io/projected/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-kube-api-access-w85j6\") pod \"horizon-7fb966687c-gnkj8\" (UID: \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\") " pod="openstack/horizon-7fb966687c-gnkj8" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.953507 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-logs\") pod \"horizon-7fb966687c-gnkj8\" (UID: \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\") " pod="openstack/horizon-7fb966687c-gnkj8" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.953524 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-horizon-secret-key\") pod \"horizon-7fb966687c-gnkj8\" (UID: \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\") " pod="openstack/horizon-7fb966687c-gnkj8" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.953541 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924-combined-ca-bundle\") pod \"neutron-db-sync-85psh\" (UID: \"0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924\") " pod="openstack/neutron-db-sync-85psh" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.953589 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-config-data\") pod \"horizon-7fb966687c-gnkj8\" (UID: \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\") " pod="openstack/horizon-7fb966687c-gnkj8" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.977728 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924-config\") pod \"neutron-db-sync-85psh\" (UID: \"0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924\") " pod="openstack/neutron-db-sync-85psh" Oct 03 13:12:06 crc kubenswrapper[4578]: I1003 13:12:06.998586 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.004277 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.005054 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924-combined-ca-bundle\") pod \"neutron-db-sync-85psh\" (UID: \"0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924\") " pod="openstack/neutron-db-sync-85psh" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.010718 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7zk9\" (UniqueName: \"kubernetes.io/projected/0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924-kube-api-access-d7zk9\") pod \"neutron-db-sync-85psh\" (UID: \"0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924\") " pod="openstack/neutron-db-sync-85psh" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.010793 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-gv9w6"] Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.012161 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gv9w6" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.020202 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-85psh" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.026084 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.026306 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-j9scz"] Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.043983 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.044297 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.044435 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.044531 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-54nlz" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.044645 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-d26lq" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.044664 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.058531 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eacf42c3-8e6f-4304-b17b-de32359e8d76-db-sync-config-data\") pod \"barbican-db-sync-j9scz\" (UID: \"eacf42c3-8e6f-4304-b17b-de32359e8d76\") " pod="openstack/barbican-db-sync-j9scz" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.058580 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-config-data\") pod \"horizon-7fb966687c-gnkj8\" (UID: \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\") " pod="openstack/horizon-7fb966687c-gnkj8" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.058697 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eacf42c3-8e6f-4304-b17b-de32359e8d76-combined-ca-bundle\") pod \"barbican-db-sync-j9scz\" (UID: \"eacf42c3-8e6f-4304-b17b-de32359e8d76\") " pod="openstack/barbican-db-sync-j9scz" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.058797 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-scripts\") pod \"horizon-7fb966687c-gnkj8\" (UID: \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\") " pod="openstack/horizon-7fb966687c-gnkj8" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.058826 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w85j6\" (UniqueName: \"kubernetes.io/projected/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-kube-api-access-w85j6\") pod \"horizon-7fb966687c-gnkj8\" (UID: \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\") " pod="openstack/horizon-7fb966687c-gnkj8" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.058867 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9lsz\" (UniqueName: \"kubernetes.io/projected/eacf42c3-8e6f-4304-b17b-de32359e8d76-kube-api-access-j9lsz\") pod \"barbican-db-sync-j9scz\" (UID: \"eacf42c3-8e6f-4304-b17b-de32359e8d76\") " pod="openstack/barbican-db-sync-j9scz" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.058890 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-logs\") pod \"horizon-7fb966687c-gnkj8\" (UID: \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\") " pod="openstack/horizon-7fb966687c-gnkj8" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.058909 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-horizon-secret-key\") pod \"horizon-7fb966687c-gnkj8\" (UID: \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\") " pod="openstack/horizon-7fb966687c-gnkj8" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.060649 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-scripts\") pod \"horizon-7fb966687c-gnkj8\" (UID: \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\") " pod="openstack/horizon-7fb966687c-gnkj8" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.061597 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-logs\") pod \"horizon-7fb966687c-gnkj8\" (UID: \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\") " pod="openstack/horizon-7fb966687c-gnkj8" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.073303 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-horizon-secret-key\") pod \"horizon-7fb966687c-gnkj8\" (UID: \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\") " pod="openstack/horizon-7fb966687c-gnkj8" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.074229 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-config-data\") pod \"horizon-7fb966687c-gnkj8\" (UID: \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\") " pod="openstack/horizon-7fb966687c-gnkj8" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.104714 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.111287 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w85j6\" (UniqueName: \"kubernetes.io/projected/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-kube-api-access-w85j6\") pod \"horizon-7fb966687c-gnkj8\" (UID: \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\") " pod="openstack/horizon-7fb966687c-gnkj8" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.146460 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-gv9w6"] Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.155007 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fb966687c-gnkj8" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.163666 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9lsz\" (UniqueName: \"kubernetes.io/projected/eacf42c3-8e6f-4304-b17b-de32359e8d76-kube-api-access-j9lsz\") pod \"barbican-db-sync-j9scz\" (UID: \"eacf42c3-8e6f-4304-b17b-de32359e8d76\") " pod="openstack/barbican-db-sync-j9scz" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.163715 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb6557a6-8665-4a81-afec-4ccaef43b801-combined-ca-bundle\") pod \"placement-db-sync-gv9w6\" (UID: \"eb6557a6-8665-4a81-afec-4ccaef43b801\") " pod="openstack/placement-db-sync-gv9w6" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.163751 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clm4q\" (UniqueName: \"kubernetes.io/projected/e393766c-2c0e-4115-87d7-a83123c0a69a-kube-api-access-clm4q\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.163775 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.163792 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-scripts\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.163823 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb6557a6-8665-4a81-afec-4ccaef43b801-config-data\") pod \"placement-db-sync-gv9w6\" (UID: \"eb6557a6-8665-4a81-afec-4ccaef43b801\") " pod="openstack/placement-db-sync-gv9w6" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.163841 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eacf42c3-8e6f-4304-b17b-de32359e8d76-db-sync-config-data\") pod \"barbican-db-sync-j9scz\" (UID: \"eacf42c3-8e6f-4304-b17b-de32359e8d76\") " pod="openstack/barbican-db-sync-j9scz" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.163864 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.163883 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e393766c-2c0e-4115-87d7-a83123c0a69a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.163900 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb6557a6-8665-4a81-afec-4ccaef43b801-scripts\") pod \"placement-db-sync-gv9w6\" (UID: \"eb6557a6-8665-4a81-afec-4ccaef43b801\") " pod="openstack/placement-db-sync-gv9w6" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.163917 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.163950 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eacf42c3-8e6f-4304-b17b-de32359e8d76-combined-ca-bundle\") pod \"barbican-db-sync-j9scz\" (UID: \"eacf42c3-8e6f-4304-b17b-de32359e8d76\") " pod="openstack/barbican-db-sync-j9scz" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.163985 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hph4p\" (UniqueName: \"kubernetes.io/projected/eb6557a6-8665-4a81-afec-4ccaef43b801-kube-api-access-hph4p\") pod \"placement-db-sync-gv9w6\" (UID: \"eb6557a6-8665-4a81-afec-4ccaef43b801\") " pod="openstack/placement-db-sync-gv9w6" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.164004 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb6557a6-8665-4a81-afec-4ccaef43b801-logs\") pod \"placement-db-sync-gv9w6\" (UID: \"eb6557a6-8665-4a81-afec-4ccaef43b801\") " pod="openstack/placement-db-sync-gv9w6" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.164027 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e393766c-2c0e-4115-87d7-a83123c0a69a-logs\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.164044 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-config-data\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.174869 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q"] Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.179773 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eacf42c3-8e6f-4304-b17b-de32359e8d76-combined-ca-bundle\") pod \"barbican-db-sync-j9scz\" (UID: \"eacf42c3-8e6f-4304-b17b-de32359e8d76\") " pod="openstack/barbican-db-sync-j9scz" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.180229 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eacf42c3-8e6f-4304-b17b-de32359e8d76-db-sync-config-data\") pod \"barbican-db-sync-j9scz\" (UID: \"eacf42c3-8e6f-4304-b17b-de32359e8d76\") " pod="openstack/barbican-db-sync-j9scz" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.266087 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e393766c-2c0e-4115-87d7-a83123c0a69a-logs\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.266128 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-config-data\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.266165 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb6557a6-8665-4a81-afec-4ccaef43b801-combined-ca-bundle\") pod \"placement-db-sync-gv9w6\" (UID: \"eb6557a6-8665-4a81-afec-4ccaef43b801\") " pod="openstack/placement-db-sync-gv9w6" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.266196 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clm4q\" (UniqueName: \"kubernetes.io/projected/e393766c-2c0e-4115-87d7-a83123c0a69a-kube-api-access-clm4q\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.266222 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.266242 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-scripts\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.266269 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb6557a6-8665-4a81-afec-4ccaef43b801-config-data\") pod \"placement-db-sync-gv9w6\" (UID: \"eb6557a6-8665-4a81-afec-4ccaef43b801\") " pod="openstack/placement-db-sync-gv9w6" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.266293 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.266313 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e393766c-2c0e-4115-87d7-a83123c0a69a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.266327 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb6557a6-8665-4a81-afec-4ccaef43b801-scripts\") pod \"placement-db-sync-gv9w6\" (UID: \"eb6557a6-8665-4a81-afec-4ccaef43b801\") " pod="openstack/placement-db-sync-gv9w6" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.266344 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.266497 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hph4p\" (UniqueName: \"kubernetes.io/projected/eb6557a6-8665-4a81-afec-4ccaef43b801-kube-api-access-hph4p\") pod \"placement-db-sync-gv9w6\" (UID: \"eb6557a6-8665-4a81-afec-4ccaef43b801\") " pod="openstack/placement-db-sync-gv9w6" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.266520 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb6557a6-8665-4a81-afec-4ccaef43b801-logs\") pod \"placement-db-sync-gv9w6\" (UID: \"eb6557a6-8665-4a81-afec-4ccaef43b801\") " pod="openstack/placement-db-sync-gv9w6" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.266715 4578 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.266947 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb6557a6-8665-4a81-afec-4ccaef43b801-logs\") pod \"placement-db-sync-gv9w6\" (UID: \"eb6557a6-8665-4a81-afec-4ccaef43b801\") " pod="openstack/placement-db-sync-gv9w6" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.269722 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e393766c-2c0e-4115-87d7-a83123c0a69a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.292471 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.292829 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9lsz\" (UniqueName: \"kubernetes.io/projected/eacf42c3-8e6f-4304-b17b-de32359e8d76-kube-api-access-j9lsz\") pod \"barbican-db-sync-j9scz\" (UID: \"eacf42c3-8e6f-4304-b17b-de32359e8d76\") " pod="openstack/barbican-db-sync-j9scz" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.293089 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e393766c-2c0e-4115-87d7-a83123c0a69a-logs\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.369342 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-2gwvm"] Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.370782 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.372918 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-j9scz" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.383802 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-2gwvm\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.391878 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-config\") pod \"dnsmasq-dns-8b5c85b87-2gwvm\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.391914 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-698hz\" (UniqueName: \"kubernetes.io/projected/3579029a-5cd2-43d7-9611-7e77a76acd5c-kube-api-access-698hz\") pod \"dnsmasq-dns-8b5c85b87-2gwvm\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.392067 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-2gwvm\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.392180 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-2gwvm\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.392307 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-2gwvm\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.387985 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb6557a6-8665-4a81-afec-4ccaef43b801-scripts\") pod \"placement-db-sync-gv9w6\" (UID: \"eb6557a6-8665-4a81-afec-4ccaef43b801\") " pod="openstack/placement-db-sync-gv9w6" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.390686 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-scripts\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.384301 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb6557a6-8665-4a81-afec-4ccaef43b801-config-data\") pod \"placement-db-sync-gv9w6\" (UID: \"eb6557a6-8665-4a81-afec-4ccaef43b801\") " pod="openstack/placement-db-sync-gv9w6" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.385743 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clm4q\" (UniqueName: \"kubernetes.io/projected/e393766c-2c0e-4115-87d7-a83123c0a69a-kube-api-access-clm4q\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.392842 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-config-data\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.416720 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb6557a6-8665-4a81-afec-4ccaef43b801-combined-ca-bundle\") pod \"placement-db-sync-gv9w6\" (UID: \"eb6557a6-8665-4a81-afec-4ccaef43b801\") " pod="openstack/placement-db-sync-gv9w6" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.419697 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.445016 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.453802 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hph4p\" (UniqueName: \"kubernetes.io/projected/eb6557a6-8665-4a81-afec-4ccaef43b801-kube-api-access-hph4p\") pod \"placement-db-sync-gv9w6\" (UID: \"eb6557a6-8665-4a81-afec-4ccaef43b801\") " pod="openstack/placement-db-sync-gv9w6" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.496232 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-2gwvm\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.496296 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-2gwvm\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.496330 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-698hz\" (UniqueName: \"kubernetes.io/projected/3579029a-5cd2-43d7-9611-7e77a76acd5c-kube-api-access-698hz\") pod \"dnsmasq-dns-8b5c85b87-2gwvm\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.496348 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-config\") pod \"dnsmasq-dns-8b5c85b87-2gwvm\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.496399 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-2gwvm\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.496454 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-2gwvm\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.497281 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-2gwvm\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.497816 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-2gwvm\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.498293 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-2gwvm\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.499494 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-2gwvm\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.501075 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-2gwvm"] Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.506434 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-config\") pod \"dnsmasq-dns-8b5c85b87-2gwvm\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.523702 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.525297 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.534597 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.535816 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.536001 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.592778 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-698hz\" (UniqueName: \"kubernetes.io/projected/3579029a-5cd2-43d7-9611-7e77a76acd5c-kube-api-access-698hz\") pod \"dnsmasq-dns-8b5c85b87-2gwvm\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.615111 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q"] Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.653151 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: E1003 13:12:07.697951 4578 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Oct 03 13:12:07 crc kubenswrapper[4578]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/5f1301ed-9d0f-4043-a967-2718ce351042/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 03 13:12:07 crc kubenswrapper[4578]: > podSandboxID="f50d837f103d8aae29adaed2d507f4d788c540cc67584a6f62e6ff1075ee03e5" Oct 03 13:12:07 crc kubenswrapper[4578]: E1003 13:12:07.698108 4578 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 03 13:12:07 crc kubenswrapper[4578]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n64h68fh95h595h67fh597hfch57ch68fh5ffh6hf4h689h659h569h65bh67bh65dh594h64h5d6hd8h5bfh9fh5c4h676h5cdh56h8bh569h664h645q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-swift-storage-0,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-swift-storage-0,SubPath:dns-swift-storage-0,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vn9pt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7ff5475cc9-8k8j7_openstack(5f1301ed-9d0f-4043-a967-2718ce351042): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/5f1301ed-9d0f-4043-a967-2718ce351042/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 03 13:12:07 crc kubenswrapper[4578]: > logger="UnhandledError" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.699993 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.700075 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.700170 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/065e7c1e-4236-4166-a21a-8a17f401f3a7-logs\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.700216 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.700237 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.700264 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thhb8\" (UniqueName: \"kubernetes.io/projected/065e7c1e-4236-4166-a21a-8a17f401f3a7-kube-api-access-thhb8\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.700300 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.700327 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/065e7c1e-4236-4166-a21a-8a17f401f3a7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: E1003 13:12:07.700406 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/5f1301ed-9d0f-4043-a967-2718ce351042/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" podUID="5f1301ed-9d0f-4043-a967-2718ce351042" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.714987 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gv9w6" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.759055 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.814158 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thhb8\" (UniqueName: \"kubernetes.io/projected/065e7c1e-4236-4166-a21a-8a17f401f3a7-kube-api-access-thhb8\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.814215 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.814237 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/065e7c1e-4236-4166-a21a-8a17f401f3a7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.814268 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.814313 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.814397 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/065e7c1e-4236-4166-a21a-8a17f401f3a7-logs\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.814437 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.814453 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.816046 4578 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.847538 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.848588 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/065e7c1e-4236-4166-a21a-8a17f401f3a7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.868318 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/065e7c1e-4236-4166-a21a-8a17f401f3a7-logs\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.894162 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.894413 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.904599 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.915274 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thhb8\" (UniqueName: \"kubernetes.io/projected/065e7c1e-4236-4166-a21a-8a17f401f3a7-kube-api-access-thhb8\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.932151 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-zmn6h"] Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.934008 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" event={"ID":"46649c43-396c-49fc-a0be-982d1805a483","Type":"ContainerStarted","Data":"45045041be5476e96ec16ba99d575f4bc08b2cd2b5ec1a507bcbe74593ba14ab"} Oct 03 13:12:07 crc kubenswrapper[4578]: I1003 13:12:07.938050 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.176842 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.361742 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-q6ppj"] Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.381004 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-85psh"] Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.395662 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c684ff4d9-2h9cq"] Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.406122 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.569410 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.597338 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-j9scz"] Oct 03 13:12:08 crc kubenswrapper[4578]: W1003 13:12:08.637041 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeacf42c3_8e6f_4304_b17b_de32359e8d76.slice/crio-e8a11b5a71783ef74f5b46f522b223afb8469a4e382eeb0af5f7b109d4090034 WatchSource:0}: Error finding container e8a11b5a71783ef74f5b46f522b223afb8469a4e382eeb0af5f7b109d4090034: Status 404 returned error can't find the container with id e8a11b5a71783ef74f5b46f522b223afb8469a4e382eeb0af5f7b109d4090034 Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.644027 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7fb966687c-gnkj8"] Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.661947 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-dns-swift-storage-0\") pod \"5f1301ed-9d0f-4043-a967-2718ce351042\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.663488 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-ovsdbserver-nb\") pod \"5f1301ed-9d0f-4043-a967-2718ce351042\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.663855 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-dns-svc\") pod \"5f1301ed-9d0f-4043-a967-2718ce351042\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.664004 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vn9pt\" (UniqueName: \"kubernetes.io/projected/5f1301ed-9d0f-4043-a967-2718ce351042-kube-api-access-vn9pt\") pod \"5f1301ed-9d0f-4043-a967-2718ce351042\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.666735 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-config\") pod \"5f1301ed-9d0f-4043-a967-2718ce351042\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.666928 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-ovsdbserver-sb\") pod \"5f1301ed-9d0f-4043-a967-2718ce351042\" (UID: \"5f1301ed-9d0f-4043-a967-2718ce351042\") " Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.671335 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f1301ed-9d0f-4043-a967-2718ce351042-kube-api-access-vn9pt" (OuterVolumeSpecName: "kube-api-access-vn9pt") pod "5f1301ed-9d0f-4043-a967-2718ce351042" (UID: "5f1301ed-9d0f-4043-a967-2718ce351042"). InnerVolumeSpecName "kube-api-access-vn9pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.770928 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vn9pt\" (UniqueName: \"kubernetes.io/projected/5f1301ed-9d0f-4043-a967-2718ce351042-kube-api-access-vn9pt\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.833367 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.891470 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5f1301ed-9d0f-4043-a967-2718ce351042" (UID: "5f1301ed-9d0f-4043-a967-2718ce351042"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.924586 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5f1301ed-9d0f-4043-a967-2718ce351042" (UID: "5f1301ed-9d0f-4043-a967-2718ce351042"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.931229 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5f1301ed-9d0f-4043-a967-2718ce351042" (UID: "5f1301ed-9d0f-4043-a967-2718ce351042"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.946168 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5f1301ed-9d0f-4043-a967-2718ce351042" (UID: "5f1301ed-9d0f-4043-a967-2718ce351042"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.977190 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.977224 4578 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.977233 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.977240 4578 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:08 crc kubenswrapper[4578]: I1003 13:12:08.994398 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-config" (OuterVolumeSpecName: "config") pod "5f1301ed-9d0f-4043-a967-2718ce351042" (UID: "5f1301ed-9d0f-4043-a967-2718ce351042"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.005947 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-gv9w6"] Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.036006 4578 generic.go:334] "Generic (PLEG): container finished" podID="46649c43-396c-49fc-a0be-982d1805a483" containerID="8597f492660039efa374b70110226fd0bfc7081ced04dfcb27c762d72968e841" exitCode=0 Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.036231 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" event={"ID":"46649c43-396c-49fc-a0be-982d1805a483","Type":"ContainerDied","Data":"8597f492660039efa374b70110226fd0bfc7081ced04dfcb27c762d72968e841"} Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.039416 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e393766c-2c0e-4115-87d7-a83123c0a69a","Type":"ContainerStarted","Data":"acf7684ff1c88a069f30560cd1e8bb9a5556c1db9a3eedf6c14e015b35bb3dcf"} Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.044069 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-j9scz" event={"ID":"eacf42c3-8e6f-4304-b17b-de32359e8d76","Type":"ContainerStarted","Data":"e8a11b5a71783ef74f5b46f522b223afb8469a4e382eeb0af5f7b109d4090034"} Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.054064 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-2gwvm"] Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.066117 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c684ff4d9-2h9cq" event={"ID":"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b","Type":"ContainerStarted","Data":"14f3997b5a24d7cbcf43bc79436222ffecd03e2e5d33f7a0cd5061909e4ea151"} Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.070747 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-85psh" event={"ID":"0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924","Type":"ContainerStarted","Data":"f0b8a13c8daf999d9420867555f31332de21fc270e85dae2adefe24914991d27"} Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.080768 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f1301ed-9d0f-4043-a967-2718ce351042-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.080861 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-q6ppj" event={"ID":"6b310034-f02f-48ac-a47a-4f41f15379a5","Type":"ContainerStarted","Data":"f4056975c2c754b545287c210946a5435f841c2b8b9c3ebb99ad66237cb15481"} Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.115070 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-85psh" podStartSLOduration=3.115043503 podStartE2EDuration="3.115043503s" podCreationTimestamp="2025-10-03 13:12:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:12:09.091990467 +0000 UTC m=+1264.890462651" watchObservedRunningTime="2025-10-03 13:12:09.115043503 +0000 UTC m=+1264.913515687" Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.122012 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zmn6h" event={"ID":"3de77cab-fde4-438d-956c-8381b6d003c2","Type":"ContainerStarted","Data":"054c0e5021dfccfa10b4b49660eacf259766addadd0d456e3e6738b69fd8d4ef"} Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.122064 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zmn6h" event={"ID":"3de77cab-fde4-438d-956c-8381b6d003c2","Type":"ContainerStarted","Data":"ec09de033df5e0e07d1133685c9fe4da3d381f833d228b30fee0d3217d09a688"} Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.125678 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6228c849-5824-4a21-8318-754c65d79a2e","Type":"ContainerStarted","Data":"848c98d057abb33e3085916996b02be877d7016c169a653c7e1a52f0bf10b9f2"} Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.126199 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.147556 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fb966687c-gnkj8" event={"ID":"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6","Type":"ContainerStarted","Data":"0b8ca28b2f9ee9e2d75784f45e79422bff81fec27b015a1eaf128c326e8f2a82"} Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.163258 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-zmn6h" podStartSLOduration=4.163238684 podStartE2EDuration="4.163238684s" podCreationTimestamp="2025-10-03 13:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:12:09.162942004 +0000 UTC m=+1264.961414188" watchObservedRunningTime="2025-10-03 13:12:09.163238684 +0000 UTC m=+1264.961710868" Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.205082 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" event={"ID":"5f1301ed-9d0f-4043-a967-2718ce351042","Type":"ContainerDied","Data":"f50d837f103d8aae29adaed2d507f4d788c540cc67584a6f62e6ff1075ee03e5"} Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.205167 4578 scope.go:117] "RemoveContainer" containerID="d796bd865b5be852491ecd46807dc296d85ff5a178c8fb01953ee842127af9c5" Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.208803 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-8k8j7" Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.358914 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-8k8j7"] Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.369264 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-8k8j7"] Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.616573 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.691371 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-dns-svc\") pod \"46649c43-396c-49fc-a0be-982d1805a483\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.691467 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-ovsdbserver-nb\") pod \"46649c43-396c-49fc-a0be-982d1805a483\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.691557 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-ovsdbserver-sb\") pod \"46649c43-396c-49fc-a0be-982d1805a483\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.692025 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-dns-swift-storage-0\") pod \"46649c43-396c-49fc-a0be-982d1805a483\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.692054 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-config\") pod \"46649c43-396c-49fc-a0be-982d1805a483\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.692182 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2p9s\" (UniqueName: \"kubernetes.io/projected/46649c43-396c-49fc-a0be-982d1805a483-kube-api-access-d2p9s\") pod \"46649c43-396c-49fc-a0be-982d1805a483\" (UID: \"46649c43-396c-49fc-a0be-982d1805a483\") " Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.718803 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46649c43-396c-49fc-a0be-982d1805a483-kube-api-access-d2p9s" (OuterVolumeSpecName: "kube-api-access-d2p9s") pod "46649c43-396c-49fc-a0be-982d1805a483" (UID: "46649c43-396c-49fc-a0be-982d1805a483"). InnerVolumeSpecName "kube-api-access-d2p9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.742842 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-config" (OuterVolumeSpecName: "config") pod "46649c43-396c-49fc-a0be-982d1805a483" (UID: "46649c43-396c-49fc-a0be-982d1805a483"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.743542 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "46649c43-396c-49fc-a0be-982d1805a483" (UID: "46649c43-396c-49fc-a0be-982d1805a483"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.754265 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "46649c43-396c-49fc-a0be-982d1805a483" (UID: "46649c43-396c-49fc-a0be-982d1805a483"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.781558 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "46649c43-396c-49fc-a0be-982d1805a483" (UID: "46649c43-396c-49fc-a0be-982d1805a483"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.790405 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "46649c43-396c-49fc-a0be-982d1805a483" (UID: "46649c43-396c-49fc-a0be-982d1805a483"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.794971 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2p9s\" (UniqueName: \"kubernetes.io/projected/46649c43-396c-49fc-a0be-982d1805a483-kube-api-access-d2p9s\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.795023 4578 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.795041 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.795052 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.795083 4578 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:09 crc kubenswrapper[4578]: I1003 13:12:09.795096 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46649c43-396c-49fc-a0be-982d1805a483-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.134784 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.193278 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c684ff4d9-2h9cq"] Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.271493 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-d5f486769-rxsds"] Oct 03 13:12:10 crc kubenswrapper[4578]: E1003 13:12:10.272438 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f1301ed-9d0f-4043-a967-2718ce351042" containerName="init" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.272451 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f1301ed-9d0f-4043-a967-2718ce351042" containerName="init" Oct 03 13:12:10 crc kubenswrapper[4578]: E1003 13:12:10.272481 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46649c43-396c-49fc-a0be-982d1805a483" containerName="init" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.272491 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="46649c43-396c-49fc-a0be-982d1805a483" containerName="init" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.272998 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="46649c43-396c-49fc-a0be-982d1805a483" containerName="init" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.273016 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f1301ed-9d0f-4043-a967-2718ce351042" containerName="init" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.274652 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d5f486769-rxsds" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.300874 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-85psh" event={"ID":"0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924","Type":"ContainerStarted","Data":"0abf83735c6600eb12b2aad4c07c5128d9e4fed23e86e7fec2aadf575ac91bcb"} Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.320103 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.322273 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a567645b-5ae1-46a0-86b1-e24b1592e060-logs\") pod \"horizon-d5f486769-rxsds\" (UID: \"a567645b-5ae1-46a0-86b1-e24b1592e060\") " pod="openstack/horizon-d5f486769-rxsds" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.329147 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a567645b-5ae1-46a0-86b1-e24b1592e060-horizon-secret-key\") pod \"horizon-d5f486769-rxsds\" (UID: \"a567645b-5ae1-46a0-86b1-e24b1592e060\") " pod="openstack/horizon-d5f486769-rxsds" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.330006 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv5wc\" (UniqueName: \"kubernetes.io/projected/a567645b-5ae1-46a0-86b1-e24b1592e060-kube-api-access-dv5wc\") pod \"horizon-d5f486769-rxsds\" (UID: \"a567645b-5ae1-46a0-86b1-e24b1592e060\") " pod="openstack/horizon-d5f486769-rxsds" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.330122 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a567645b-5ae1-46a0-86b1-e24b1592e060-config-data\") pod \"horizon-d5f486769-rxsds\" (UID: \"a567645b-5ae1-46a0-86b1-e24b1592e060\") " pod="openstack/horizon-d5f486769-rxsds" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.330261 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a567645b-5ae1-46a0-86b1-e24b1592e060-scripts\") pod \"horizon-d5f486769-rxsds\" (UID: \"a567645b-5ae1-46a0-86b1-e24b1592e060\") " pod="openstack/horizon-d5f486769-rxsds" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.346568 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.371118 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"065e7c1e-4236-4166-a21a-8a17f401f3a7","Type":"ContainerStarted","Data":"df391433ac060f84b7d6eddc5fd184d7a388b853e26636ebf167344a9694a3d9"} Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.381131 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d5f486769-rxsds"] Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.405014 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gv9w6" event={"ID":"eb6557a6-8665-4a81-afec-4ccaef43b801","Type":"ContainerStarted","Data":"65ea7c7a946a2329012a28d8b4a6426be4cbb51ee7f3f5028c004af036781f12"} Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.412612 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" event={"ID":"46649c43-396c-49fc-a0be-982d1805a483","Type":"ContainerDied","Data":"45045041be5476e96ec16ba99d575f4bc08b2cd2b5ec1a507bcbe74593ba14ab"} Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.412737 4578 scope.go:117] "RemoveContainer" containerID="8597f492660039efa374b70110226fd0bfc7081ced04dfcb27c762d72968e841" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.412904 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.433413 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a567645b-5ae1-46a0-86b1-e24b1592e060-logs\") pod \"horizon-d5f486769-rxsds\" (UID: \"a567645b-5ae1-46a0-86b1-e24b1592e060\") " pod="openstack/horizon-d5f486769-rxsds" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.434149 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a567645b-5ae1-46a0-86b1-e24b1592e060-horizon-secret-key\") pod \"horizon-d5f486769-rxsds\" (UID: \"a567645b-5ae1-46a0-86b1-e24b1592e060\") " pod="openstack/horizon-d5f486769-rxsds" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.434310 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv5wc\" (UniqueName: \"kubernetes.io/projected/a567645b-5ae1-46a0-86b1-e24b1592e060-kube-api-access-dv5wc\") pod \"horizon-d5f486769-rxsds\" (UID: \"a567645b-5ae1-46a0-86b1-e24b1592e060\") " pod="openstack/horizon-d5f486769-rxsds" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.434387 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a567645b-5ae1-46a0-86b1-e24b1592e060-config-data\") pod \"horizon-d5f486769-rxsds\" (UID: \"a567645b-5ae1-46a0-86b1-e24b1592e060\") " pod="openstack/horizon-d5f486769-rxsds" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.434513 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a567645b-5ae1-46a0-86b1-e24b1592e060-scripts\") pod \"horizon-d5f486769-rxsds\" (UID: \"a567645b-5ae1-46a0-86b1-e24b1592e060\") " pod="openstack/horizon-d5f486769-rxsds" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.434952 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a567645b-5ae1-46a0-86b1-e24b1592e060-logs\") pod \"horizon-d5f486769-rxsds\" (UID: \"a567645b-5ae1-46a0-86b1-e24b1592e060\") " pod="openstack/horizon-d5f486769-rxsds" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.443380 4578 generic.go:334] "Generic (PLEG): container finished" podID="3579029a-5cd2-43d7-9611-7e77a76acd5c" containerID="4239f81a6442d31df07d455e6472259d868ed6c5074d66df49a2660b6459b68d" exitCode=0 Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.443472 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" event={"ID":"3579029a-5cd2-43d7-9611-7e77a76acd5c","Type":"ContainerDied","Data":"4239f81a6442d31df07d455e6472259d868ed6c5074d66df49a2660b6459b68d"} Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.443499 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" event={"ID":"3579029a-5cd2-43d7-9611-7e77a76acd5c","Type":"ContainerStarted","Data":"ebd9e191f588b9275e570c4edd5552be6859879c0043efe903f921a6400718c7"} Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.459142 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a567645b-5ae1-46a0-86b1-e24b1592e060-scripts\") pod \"horizon-d5f486769-rxsds\" (UID: \"a567645b-5ae1-46a0-86b1-e24b1592e060\") " pod="openstack/horizon-d5f486769-rxsds" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.460587 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a567645b-5ae1-46a0-86b1-e24b1592e060-config-data\") pod \"horizon-d5f486769-rxsds\" (UID: \"a567645b-5ae1-46a0-86b1-e24b1592e060\") " pod="openstack/horizon-d5f486769-rxsds" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.468068 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a567645b-5ae1-46a0-86b1-e24b1592e060-horizon-secret-key\") pod \"horizon-d5f486769-rxsds\" (UID: \"a567645b-5ae1-46a0-86b1-e24b1592e060\") " pod="openstack/horizon-d5f486769-rxsds" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.511688 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q"] Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.523497 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-qlj9q"] Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.525225 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv5wc\" (UniqueName: \"kubernetes.io/projected/a567645b-5ae1-46a0-86b1-e24b1592e060-kube-api-access-dv5wc\") pod \"horizon-d5f486769-rxsds\" (UID: \"a567645b-5ae1-46a0-86b1-e24b1592e060\") " pod="openstack/horizon-d5f486769-rxsds" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.629045 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d5f486769-rxsds" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.944729 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46649c43-396c-49fc-a0be-982d1805a483" path="/var/lib/kubelet/pods/46649c43-396c-49fc-a0be-982d1805a483/volumes" Oct 03 13:12:10 crc kubenswrapper[4578]: I1003 13:12:10.946112 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f1301ed-9d0f-4043-a967-2718ce351042" path="/var/lib/kubelet/pods/5f1301ed-9d0f-4043-a967-2718ce351042/volumes" Oct 03 13:12:11 crc kubenswrapper[4578]: I1003 13:12:11.317175 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d5f486769-rxsds"] Oct 03 13:12:11 crc kubenswrapper[4578]: I1003 13:12:11.522719 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" event={"ID":"3579029a-5cd2-43d7-9611-7e77a76acd5c","Type":"ContainerStarted","Data":"5e3db199e76de0240c635a99e2ecc1c9e01f4c85de1237bb45f06ad520a825eb"} Oct 03 13:12:11 crc kubenswrapper[4578]: I1003 13:12:11.522952 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:11 crc kubenswrapper[4578]: I1003 13:12:11.553137 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e393766c-2c0e-4115-87d7-a83123c0a69a","Type":"ContainerStarted","Data":"85b3db0158584d61f4f703980ae36fcf5f65c73c9a4aa65bcf729017f317ce18"} Oct 03 13:12:11 crc kubenswrapper[4578]: I1003 13:12:11.554775 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d5f486769-rxsds" event={"ID":"a567645b-5ae1-46a0-86b1-e24b1592e060","Type":"ContainerStarted","Data":"12363d6128e3c56437080d84701c50a13a9ab3b59cfcfd2c71de90be08a13a2b"} Oct 03 13:12:11 crc kubenswrapper[4578]: I1003 13:12:11.569441 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"065e7c1e-4236-4166-a21a-8a17f401f3a7","Type":"ContainerStarted","Data":"21025c71f4960a946df5aaa1356ed6f7a6c4dde1fdadd130ae5338921dc98a46"} Oct 03 13:12:12 crc kubenswrapper[4578]: I1003 13:12:12.586360 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"065e7c1e-4236-4166-a21a-8a17f401f3a7","Type":"ContainerStarted","Data":"45d6d13ae99319b05cdef5a42d66a25f545a3e3b4ce5edb4a9e539905feb3b65"} Oct 03 13:12:12 crc kubenswrapper[4578]: I1003 13:12:12.586765 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="065e7c1e-4236-4166-a21a-8a17f401f3a7" containerName="glance-log" containerID="cri-o://21025c71f4960a946df5aaa1356ed6f7a6c4dde1fdadd130ae5338921dc98a46" gracePeriod=30 Oct 03 13:12:12 crc kubenswrapper[4578]: I1003 13:12:12.586770 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="065e7c1e-4236-4166-a21a-8a17f401f3a7" containerName="glance-httpd" containerID="cri-o://45d6d13ae99319b05cdef5a42d66a25f545a3e3b4ce5edb4a9e539905feb3b65" gracePeriod=30 Oct 03 13:12:12 crc kubenswrapper[4578]: I1003 13:12:12.613860 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.613836436 podStartE2EDuration="6.613836436s" podCreationTimestamp="2025-10-03 13:12:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:12:12.609333024 +0000 UTC m=+1268.407805208" watchObservedRunningTime="2025-10-03 13:12:12.613836436 +0000 UTC m=+1268.412308620" Oct 03 13:12:12 crc kubenswrapper[4578]: I1003 13:12:12.615037 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" podStartSLOduration=5.615024984 podStartE2EDuration="5.615024984s" podCreationTimestamp="2025-10-03 13:12:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:12:11.550204168 +0000 UTC m=+1267.348676352" watchObservedRunningTime="2025-10-03 13:12:12.615024984 +0000 UTC m=+1268.413497168" Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.618330 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e393766c-2c0e-4115-87d7-a83123c0a69a","Type":"ContainerStarted","Data":"fde74a6d3b91a522363a3d8c034ed163efc96dff7b22df422dbbb6fc3dc5d193"} Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.618382 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e393766c-2c0e-4115-87d7-a83123c0a69a" containerName="glance-log" containerID="cri-o://85b3db0158584d61f4f703980ae36fcf5f65c73c9a4aa65bcf729017f317ce18" gracePeriod=30 Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.618807 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e393766c-2c0e-4115-87d7-a83123c0a69a" containerName="glance-httpd" containerID="cri-o://fde74a6d3b91a522363a3d8c034ed163efc96dff7b22df422dbbb6fc3dc5d193" gracePeriod=30 Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.636926 4578 generic.go:334] "Generic (PLEG): container finished" podID="065e7c1e-4236-4166-a21a-8a17f401f3a7" containerID="45d6d13ae99319b05cdef5a42d66a25f545a3e3b4ce5edb4a9e539905feb3b65" exitCode=143 Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.636989 4578 generic.go:334] "Generic (PLEG): container finished" podID="065e7c1e-4236-4166-a21a-8a17f401f3a7" containerID="21025c71f4960a946df5aaa1356ed6f7a6c4dde1fdadd130ae5338921dc98a46" exitCode=143 Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.637013 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"065e7c1e-4236-4166-a21a-8a17f401f3a7","Type":"ContainerDied","Data":"45d6d13ae99319b05cdef5a42d66a25f545a3e3b4ce5edb4a9e539905feb3b65"} Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.637040 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"065e7c1e-4236-4166-a21a-8a17f401f3a7","Type":"ContainerDied","Data":"21025c71f4960a946df5aaa1356ed6f7a6c4dde1fdadd130ae5338921dc98a46"} Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.667353 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.667321452 podStartE2EDuration="7.667321452s" podCreationTimestamp="2025-10-03 13:12:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:12:13.643230023 +0000 UTC m=+1269.441702227" watchObservedRunningTime="2025-10-03 13:12:13.667321452 +0000 UTC m=+1269.465793636" Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.684605 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.723580 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-internal-tls-certs\") pod \"065e7c1e-4236-4166-a21a-8a17f401f3a7\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.725928 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/065e7c1e-4236-4166-a21a-8a17f401f3a7-logs\") pod \"065e7c1e-4236-4166-a21a-8a17f401f3a7\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.726085 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-combined-ca-bundle\") pod \"065e7c1e-4236-4166-a21a-8a17f401f3a7\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.726119 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"065e7c1e-4236-4166-a21a-8a17f401f3a7\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.726174 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-config-data\") pod \"065e7c1e-4236-4166-a21a-8a17f401f3a7\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.726219 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-scripts\") pod \"065e7c1e-4236-4166-a21a-8a17f401f3a7\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.726255 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thhb8\" (UniqueName: \"kubernetes.io/projected/065e7c1e-4236-4166-a21a-8a17f401f3a7-kube-api-access-thhb8\") pod \"065e7c1e-4236-4166-a21a-8a17f401f3a7\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.726292 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/065e7c1e-4236-4166-a21a-8a17f401f3a7-httpd-run\") pod \"065e7c1e-4236-4166-a21a-8a17f401f3a7\" (UID: \"065e7c1e-4236-4166-a21a-8a17f401f3a7\") " Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.726650 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/065e7c1e-4236-4166-a21a-8a17f401f3a7-logs" (OuterVolumeSpecName: "logs") pod "065e7c1e-4236-4166-a21a-8a17f401f3a7" (UID: "065e7c1e-4236-4166-a21a-8a17f401f3a7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.727131 4578 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/065e7c1e-4236-4166-a21a-8a17f401f3a7-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.727289 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/065e7c1e-4236-4166-a21a-8a17f401f3a7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "065e7c1e-4236-4166-a21a-8a17f401f3a7" (UID: "065e7c1e-4236-4166-a21a-8a17f401f3a7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.756888 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/065e7c1e-4236-4166-a21a-8a17f401f3a7-kube-api-access-thhb8" (OuterVolumeSpecName: "kube-api-access-thhb8") pod "065e7c1e-4236-4166-a21a-8a17f401f3a7" (UID: "065e7c1e-4236-4166-a21a-8a17f401f3a7"). InnerVolumeSpecName "kube-api-access-thhb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.761939 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "065e7c1e-4236-4166-a21a-8a17f401f3a7" (UID: "065e7c1e-4236-4166-a21a-8a17f401f3a7"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.767126 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-scripts" (OuterVolumeSpecName: "scripts") pod "065e7c1e-4236-4166-a21a-8a17f401f3a7" (UID: "065e7c1e-4236-4166-a21a-8a17f401f3a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.801698 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "065e7c1e-4236-4166-a21a-8a17f401f3a7" (UID: "065e7c1e-4236-4166-a21a-8a17f401f3a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.828530 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.828595 4578 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.828608 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.828621 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thhb8\" (UniqueName: \"kubernetes.io/projected/065e7c1e-4236-4166-a21a-8a17f401f3a7-kube-api-access-thhb8\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.828663 4578 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/065e7c1e-4236-4166-a21a-8a17f401f3a7-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.846553 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-config-data" (OuterVolumeSpecName: "config-data") pod "065e7c1e-4236-4166-a21a-8a17f401f3a7" (UID: "065e7c1e-4236-4166-a21a-8a17f401f3a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.854223 4578 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.903825 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "065e7c1e-4236-4166-a21a-8a17f401f3a7" (UID: "065e7c1e-4236-4166-a21a-8a17f401f3a7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.933955 4578 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.933985 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:13 crc kubenswrapper[4578]: I1003 13:12:13.933996 4578 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/065e7c1e-4236-4166-a21a-8a17f401f3a7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.405907 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.543463 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"e393766c-2c0e-4115-87d7-a83123c0a69a\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.543554 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clm4q\" (UniqueName: \"kubernetes.io/projected/e393766c-2c0e-4115-87d7-a83123c0a69a-kube-api-access-clm4q\") pod \"e393766c-2c0e-4115-87d7-a83123c0a69a\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.543585 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e393766c-2c0e-4115-87d7-a83123c0a69a-httpd-run\") pod \"e393766c-2c0e-4115-87d7-a83123c0a69a\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.543645 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-combined-ca-bundle\") pod \"e393766c-2c0e-4115-87d7-a83123c0a69a\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.543730 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-public-tls-certs\") pod \"e393766c-2c0e-4115-87d7-a83123c0a69a\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.543747 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e393766c-2c0e-4115-87d7-a83123c0a69a-logs\") pod \"e393766c-2c0e-4115-87d7-a83123c0a69a\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.543777 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-scripts\") pod \"e393766c-2c0e-4115-87d7-a83123c0a69a\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.543837 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-config-data\") pod \"e393766c-2c0e-4115-87d7-a83123c0a69a\" (UID: \"e393766c-2c0e-4115-87d7-a83123c0a69a\") " Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.545674 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e393766c-2c0e-4115-87d7-a83123c0a69a-logs" (OuterVolumeSpecName: "logs") pod "e393766c-2c0e-4115-87d7-a83123c0a69a" (UID: "e393766c-2c0e-4115-87d7-a83123c0a69a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.546503 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e393766c-2c0e-4115-87d7-a83123c0a69a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e393766c-2c0e-4115-87d7-a83123c0a69a" (UID: "e393766c-2c0e-4115-87d7-a83123c0a69a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.557492 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-scripts" (OuterVolumeSpecName: "scripts") pod "e393766c-2c0e-4115-87d7-a83123c0a69a" (UID: "e393766c-2c0e-4115-87d7-a83123c0a69a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.557616 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e393766c-2c0e-4115-87d7-a83123c0a69a-kube-api-access-clm4q" (OuterVolumeSpecName: "kube-api-access-clm4q") pod "e393766c-2c0e-4115-87d7-a83123c0a69a" (UID: "e393766c-2c0e-4115-87d7-a83123c0a69a"). InnerVolumeSpecName "kube-api-access-clm4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.558185 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "e393766c-2c0e-4115-87d7-a83123c0a69a" (UID: "e393766c-2c0e-4115-87d7-a83123c0a69a"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.625878 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-config-data" (OuterVolumeSpecName: "config-data") pod "e393766c-2c0e-4115-87d7-a83123c0a69a" (UID: "e393766c-2c0e-4115-87d7-a83123c0a69a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.626133 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e393766c-2c0e-4115-87d7-a83123c0a69a" (UID: "e393766c-2c0e-4115-87d7-a83123c0a69a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.628213 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e393766c-2c0e-4115-87d7-a83123c0a69a" (UID: "e393766c-2c0e-4115-87d7-a83123c0a69a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.646169 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clm4q\" (UniqueName: \"kubernetes.io/projected/e393766c-2c0e-4115-87d7-a83123c0a69a-kube-api-access-clm4q\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.646203 4578 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e393766c-2c0e-4115-87d7-a83123c0a69a-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.646216 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.646229 4578 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.646240 4578 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e393766c-2c0e-4115-87d7-a83123c0a69a-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.646250 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.646260 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e393766c-2c0e-4115-87d7-a83123c0a69a-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.646308 4578 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.662831 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"065e7c1e-4236-4166-a21a-8a17f401f3a7","Type":"ContainerDied","Data":"df391433ac060f84b7d6eddc5fd184d7a388b853e26636ebf167344a9694a3d9"} Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.674027 4578 scope.go:117] "RemoveContainer" containerID="45d6d13ae99319b05cdef5a42d66a25f545a3e3b4ce5edb4a9e539905feb3b65" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.668849 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.677011 4578 generic.go:334] "Generic (PLEG): container finished" podID="e393766c-2c0e-4115-87d7-a83123c0a69a" containerID="fde74a6d3b91a522363a3d8c034ed163efc96dff7b22df422dbbb6fc3dc5d193" exitCode=0 Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.677039 4578 generic.go:334] "Generic (PLEG): container finished" podID="e393766c-2c0e-4115-87d7-a83123c0a69a" containerID="85b3db0158584d61f4f703980ae36fcf5f65c73c9a4aa65bcf729017f317ce18" exitCode=143 Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.677056 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e393766c-2c0e-4115-87d7-a83123c0a69a","Type":"ContainerDied","Data":"fde74a6d3b91a522363a3d8c034ed163efc96dff7b22df422dbbb6fc3dc5d193"} Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.677079 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e393766c-2c0e-4115-87d7-a83123c0a69a","Type":"ContainerDied","Data":"85b3db0158584d61f4f703980ae36fcf5f65c73c9a4aa65bcf729017f317ce18"} Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.677088 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e393766c-2c0e-4115-87d7-a83123c0a69a","Type":"ContainerDied","Data":"acf7684ff1c88a069f30560cd1e8bb9a5556c1db9a3eedf6c14e015b35bb3dcf"} Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.677285 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.699703 4578 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.741022 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.752112 4578 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.752155 4578 scope.go:117] "RemoveContainer" containerID="21025c71f4960a946df5aaa1356ed6f7a6c4dde1fdadd130ae5338921dc98a46" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.760531 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.781749 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.840795 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.855009 4578 scope.go:117] "RemoveContainer" containerID="fde74a6d3b91a522363a3d8c034ed163efc96dff7b22df422dbbb6fc3dc5d193" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.864037 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:12:14 crc kubenswrapper[4578]: E1003 13:12:14.864895 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="065e7c1e-4236-4166-a21a-8a17f401f3a7" containerName="glance-httpd" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.864916 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="065e7c1e-4236-4166-a21a-8a17f401f3a7" containerName="glance-httpd" Oct 03 13:12:14 crc kubenswrapper[4578]: E1003 13:12:14.864931 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="065e7c1e-4236-4166-a21a-8a17f401f3a7" containerName="glance-log" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.864938 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="065e7c1e-4236-4166-a21a-8a17f401f3a7" containerName="glance-log" Oct 03 13:12:14 crc kubenswrapper[4578]: E1003 13:12:14.864976 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e393766c-2c0e-4115-87d7-a83123c0a69a" containerName="glance-log" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.864984 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="e393766c-2c0e-4115-87d7-a83123c0a69a" containerName="glance-log" Oct 03 13:12:14 crc kubenswrapper[4578]: E1003 13:12:14.864995 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e393766c-2c0e-4115-87d7-a83123c0a69a" containerName="glance-httpd" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.865003 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="e393766c-2c0e-4115-87d7-a83123c0a69a" containerName="glance-httpd" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.865297 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="e393766c-2c0e-4115-87d7-a83123c0a69a" containerName="glance-httpd" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.865321 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="065e7c1e-4236-4166-a21a-8a17f401f3a7" containerName="glance-httpd" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.865331 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="065e7c1e-4236-4166-a21a-8a17f401f3a7" containerName="glance-log" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.865355 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="e393766c-2c0e-4115-87d7-a83123c0a69a" containerName="glance-log" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.874044 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.881244 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.881397 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-54nlz" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.882396 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.885784 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.907108 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.935277 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="065e7c1e-4236-4166-a21a-8a17f401f3a7" path="/var/lib/kubelet/pods/065e7c1e-4236-4166-a21a-8a17f401f3a7/volumes" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.936084 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e393766c-2c0e-4115-87d7-a83123c0a69a" path="/var/lib/kubelet/pods/e393766c-2c0e-4115-87d7-a83123c0a69a/volumes" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.936739 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.942948 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.945482 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.946950 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.948766 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:12:14 crc kubenswrapper[4578]: I1003 13:12:14.958313 4578 scope.go:117] "RemoveContainer" containerID="85b3db0158584d61f4f703980ae36fcf5f65c73c9a4aa65bcf729017f317ce18" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.001865 4578 scope.go:117] "RemoveContainer" containerID="fde74a6d3b91a522363a3d8c034ed163efc96dff7b22df422dbbb6fc3dc5d193" Oct 03 13:12:15 crc kubenswrapper[4578]: E1003 13:12:15.005399 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fde74a6d3b91a522363a3d8c034ed163efc96dff7b22df422dbbb6fc3dc5d193\": container with ID starting with fde74a6d3b91a522363a3d8c034ed163efc96dff7b22df422dbbb6fc3dc5d193 not found: ID does not exist" containerID="fde74a6d3b91a522363a3d8c034ed163efc96dff7b22df422dbbb6fc3dc5d193" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.005440 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fde74a6d3b91a522363a3d8c034ed163efc96dff7b22df422dbbb6fc3dc5d193"} err="failed to get container status \"fde74a6d3b91a522363a3d8c034ed163efc96dff7b22df422dbbb6fc3dc5d193\": rpc error: code = NotFound desc = could not find container \"fde74a6d3b91a522363a3d8c034ed163efc96dff7b22df422dbbb6fc3dc5d193\": container with ID starting with fde74a6d3b91a522363a3d8c034ed163efc96dff7b22df422dbbb6fc3dc5d193 not found: ID does not exist" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.005466 4578 scope.go:117] "RemoveContainer" containerID="85b3db0158584d61f4f703980ae36fcf5f65c73c9a4aa65bcf729017f317ce18" Oct 03 13:12:15 crc kubenswrapper[4578]: E1003 13:12:15.006063 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85b3db0158584d61f4f703980ae36fcf5f65c73c9a4aa65bcf729017f317ce18\": container with ID starting with 85b3db0158584d61f4f703980ae36fcf5f65c73c9a4aa65bcf729017f317ce18 not found: ID does not exist" containerID="85b3db0158584d61f4f703980ae36fcf5f65c73c9a4aa65bcf729017f317ce18" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.006086 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85b3db0158584d61f4f703980ae36fcf5f65c73c9a4aa65bcf729017f317ce18"} err="failed to get container status \"85b3db0158584d61f4f703980ae36fcf5f65c73c9a4aa65bcf729017f317ce18\": rpc error: code = NotFound desc = could not find container \"85b3db0158584d61f4f703980ae36fcf5f65c73c9a4aa65bcf729017f317ce18\": container with ID starting with 85b3db0158584d61f4f703980ae36fcf5f65c73c9a4aa65bcf729017f317ce18 not found: ID does not exist" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.006103 4578 scope.go:117] "RemoveContainer" containerID="fde74a6d3b91a522363a3d8c034ed163efc96dff7b22df422dbbb6fc3dc5d193" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.010160 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fde74a6d3b91a522363a3d8c034ed163efc96dff7b22df422dbbb6fc3dc5d193"} err="failed to get container status \"fde74a6d3b91a522363a3d8c034ed163efc96dff7b22df422dbbb6fc3dc5d193\": rpc error: code = NotFound desc = could not find container \"fde74a6d3b91a522363a3d8c034ed163efc96dff7b22df422dbbb6fc3dc5d193\": container with ID starting with fde74a6d3b91a522363a3d8c034ed163efc96dff7b22df422dbbb6fc3dc5d193 not found: ID does not exist" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.010214 4578 scope.go:117] "RemoveContainer" containerID="85b3db0158584d61f4f703980ae36fcf5f65c73c9a4aa65bcf729017f317ce18" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.013938 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85b3db0158584d61f4f703980ae36fcf5f65c73c9a4aa65bcf729017f317ce18"} err="failed to get container status \"85b3db0158584d61f4f703980ae36fcf5f65c73c9a4aa65bcf729017f317ce18\": rpc error: code = NotFound desc = could not find container \"85b3db0158584d61f4f703980ae36fcf5f65c73c9a4aa65bcf729017f317ce18\": container with ID starting with 85b3db0158584d61f4f703980ae36fcf5f65c73c9a4aa65bcf729017f317ce18 not found: ID does not exist" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.069261 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfssl\" (UniqueName: \"kubernetes.io/projected/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-kube-api-access-lfssl\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.069355 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fp6v\" (UniqueName: \"kubernetes.io/projected/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-kube-api-access-4fp6v\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.069376 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.069408 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.069441 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.069509 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-logs\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.069534 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.069572 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.069593 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-scripts\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.069613 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.069658 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.069674 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-logs\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.069700 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.069714 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.069743 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.069765 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-config-data\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.171285 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.171330 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-scripts\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.171354 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.171389 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.171406 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-logs\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.171432 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.171454 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.171480 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.171496 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-config-data\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.171524 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfssl\" (UniqueName: \"kubernetes.io/projected/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-kube-api-access-lfssl\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.171551 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fp6v\" (UniqueName: \"kubernetes.io/projected/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-kube-api-access-4fp6v\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.171567 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.171593 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.171622 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.171666 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-logs\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.171695 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.171850 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.172324 4578 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.172498 4578 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.180417 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.180740 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-logs\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.181337 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.183903 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-config-data\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.185144 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.185399 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-logs\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.186130 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.183752 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-scripts\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.198202 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fp6v\" (UniqueName: \"kubernetes.io/projected/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-kube-api-access-4fp6v\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.198720 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.199098 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.199247 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.209754 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfssl\" (UniqueName: \"kubernetes.io/projected/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-kube-api-access-lfssl\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.234178 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.236199 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.287359 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.509284 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.711676 4578 generic.go:334] "Generic (PLEG): container finished" podID="3de77cab-fde4-438d-956c-8381b6d003c2" containerID="054c0e5021dfccfa10b4b49660eacf259766addadd0d456e3e6738b69fd8d4ef" exitCode=0 Oct 03 13:12:15 crc kubenswrapper[4578]: I1003 13:12:15.711758 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zmn6h" event={"ID":"3de77cab-fde4-438d-956c-8381b6d003c2","Type":"ContainerDied","Data":"054c0e5021dfccfa10b4b49660eacf259766addadd0d456e3e6738b69fd8d4ef"} Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.015405 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.369001 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7fb966687c-gnkj8"] Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.421489 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7b58cf9568-f8nng"] Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.438480 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.456682 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.512482 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.534755 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7b58cf9568-f8nng"] Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.567846 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.586783 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d5f486769-rxsds"] Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.599538 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-68684899bd-2cwll"] Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.609360 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68684899bd-2cwll"] Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.610855 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.620312 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a54aea6a-c051-4832-b764-3ef53a9cc3a1-logs\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.620381 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a54aea6a-c051-4832-b764-3ef53a9cc3a1-horizon-secret-key\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.620475 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2q9l\" (UniqueName: \"kubernetes.io/projected/a54aea6a-c051-4832-b764-3ef53a9cc3a1-kube-api-access-q2q9l\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.620505 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a54aea6a-c051-4832-b764-3ef53a9cc3a1-horizon-tls-certs\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.620687 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a54aea6a-c051-4832-b764-3ef53a9cc3a1-scripts\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.620793 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a54aea6a-c051-4832-b764-3ef53a9cc3a1-combined-ca-bundle\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.620845 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a54aea6a-c051-4832-b764-3ef53a9cc3a1-config-data\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.724717 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk5hs\" (UniqueName: \"kubernetes.io/projected/d49d623e-2cd6-4786-b1da-34b78b89a134-kube-api-access-lk5hs\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.725037 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a54aea6a-c051-4832-b764-3ef53a9cc3a1-scripts\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.725059 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a54aea6a-c051-4832-b764-3ef53a9cc3a1-combined-ca-bundle\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.725133 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a54aea6a-c051-4832-b764-3ef53a9cc3a1-config-data\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.725184 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a54aea6a-c051-4832-b764-3ef53a9cc3a1-logs\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.725208 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a54aea6a-c051-4832-b764-3ef53a9cc3a1-horizon-secret-key\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.725224 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d49d623e-2cd6-4786-b1da-34b78b89a134-config-data\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.725246 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d49d623e-2cd6-4786-b1da-34b78b89a134-scripts\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.725277 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2q9l\" (UniqueName: \"kubernetes.io/projected/a54aea6a-c051-4832-b764-3ef53a9cc3a1-kube-api-access-q2q9l\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.725298 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a54aea6a-c051-4832-b764-3ef53a9cc3a1-horizon-tls-certs\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.725319 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d49d623e-2cd6-4786-b1da-34b78b89a134-horizon-secret-key\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.725337 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d49d623e-2cd6-4786-b1da-34b78b89a134-combined-ca-bundle\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.725369 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d49d623e-2cd6-4786-b1da-34b78b89a134-horizon-tls-certs\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.725384 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d49d623e-2cd6-4786-b1da-34b78b89a134-logs\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.725848 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a54aea6a-c051-4832-b764-3ef53a9cc3a1-logs\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.726461 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a54aea6a-c051-4832-b764-3ef53a9cc3a1-config-data\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.727466 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a54aea6a-c051-4832-b764-3ef53a9cc3a1-scripts\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.732892 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a54aea6a-c051-4832-b764-3ef53a9cc3a1-horizon-secret-key\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.734195 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a54aea6a-c051-4832-b764-3ef53a9cc3a1-combined-ca-bundle\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.739160 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a54aea6a-c051-4832-b764-3ef53a9cc3a1-horizon-tls-certs\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.765875 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2q9l\" (UniqueName: \"kubernetes.io/projected/a54aea6a-c051-4832-b764-3ef53a9cc3a1-kube-api-access-q2q9l\") pod \"horizon-7b58cf9568-f8nng\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.779666 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.827579 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d49d623e-2cd6-4786-b1da-34b78b89a134-horizon-secret-key\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.827622 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d49d623e-2cd6-4786-b1da-34b78b89a134-combined-ca-bundle\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.827676 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d49d623e-2cd6-4786-b1da-34b78b89a134-horizon-tls-certs\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.827694 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d49d623e-2cd6-4786-b1da-34b78b89a134-logs\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.827729 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk5hs\" (UniqueName: \"kubernetes.io/projected/d49d623e-2cd6-4786-b1da-34b78b89a134-kube-api-access-lk5hs\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.827855 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d49d623e-2cd6-4786-b1da-34b78b89a134-config-data\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.827881 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d49d623e-2cd6-4786-b1da-34b78b89a134-scripts\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.828683 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d49d623e-2cd6-4786-b1da-34b78b89a134-scripts\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.829971 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d49d623e-2cd6-4786-b1da-34b78b89a134-logs\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.830703 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d49d623e-2cd6-4786-b1da-34b78b89a134-config-data\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.832563 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d49d623e-2cd6-4786-b1da-34b78b89a134-horizon-tls-certs\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.833130 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d49d623e-2cd6-4786-b1da-34b78b89a134-horizon-secret-key\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.834256 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d49d623e-2cd6-4786-b1da-34b78b89a134-combined-ca-bundle\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.847038 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk5hs\" (UniqueName: \"kubernetes.io/projected/d49d623e-2cd6-4786-b1da-34b78b89a134-kube-api-access-lk5hs\") pod \"horizon-68684899bd-2cwll\" (UID: \"d49d623e-2cd6-4786-b1da-34b78b89a134\") " pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:16 crc kubenswrapper[4578]: I1003 13:12:16.946293 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:17 crc kubenswrapper[4578]: I1003 13:12:17.760743 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:17 crc kubenswrapper[4578]: I1003 13:12:17.852373 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-s7mxj"] Oct 03 13:12:17 crc kubenswrapper[4578]: I1003 13:12:17.852595 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" podUID="92cd4ef0-355f-4564-9590-6593431b678a" containerName="dnsmasq-dns" containerID="cri-o://67a6dda3e987feb1381d7012ff7fd320626f3627c9b84978669afe0518d9e937" gracePeriod=10 Oct 03 13:12:18 crc kubenswrapper[4578]: I1003 13:12:18.759742 4578 generic.go:334] "Generic (PLEG): container finished" podID="92cd4ef0-355f-4564-9590-6593431b678a" containerID="67a6dda3e987feb1381d7012ff7fd320626f3627c9b84978669afe0518d9e937" exitCode=0 Oct 03 13:12:18 crc kubenswrapper[4578]: I1003 13:12:18.759783 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" event={"ID":"92cd4ef0-355f-4564-9590-6593431b678a","Type":"ContainerDied","Data":"67a6dda3e987feb1381d7012ff7fd320626f3627c9b84978669afe0518d9e937"} Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.339892 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.489263 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-combined-ca-bundle\") pod \"3de77cab-fde4-438d-956c-8381b6d003c2\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.489322 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-config-data\") pod \"3de77cab-fde4-438d-956c-8381b6d003c2\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.489378 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-fernet-keys\") pod \"3de77cab-fde4-438d-956c-8381b6d003c2\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.490396 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-662gl\" (UniqueName: \"kubernetes.io/projected/3de77cab-fde4-438d-956c-8381b6d003c2-kube-api-access-662gl\") pod \"3de77cab-fde4-438d-956c-8381b6d003c2\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.490449 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-credential-keys\") pod \"3de77cab-fde4-438d-956c-8381b6d003c2\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.490485 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-scripts\") pod \"3de77cab-fde4-438d-956c-8381b6d003c2\" (UID: \"3de77cab-fde4-438d-956c-8381b6d003c2\") " Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.496165 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3de77cab-fde4-438d-956c-8381b6d003c2-kube-api-access-662gl" (OuterVolumeSpecName: "kube-api-access-662gl") pod "3de77cab-fde4-438d-956c-8381b6d003c2" (UID: "3de77cab-fde4-438d-956c-8381b6d003c2"). InnerVolumeSpecName "kube-api-access-662gl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.497350 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-scripts" (OuterVolumeSpecName: "scripts") pod "3de77cab-fde4-438d-956c-8381b6d003c2" (UID: "3de77cab-fde4-438d-956c-8381b6d003c2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.497731 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "3de77cab-fde4-438d-956c-8381b6d003c2" (UID: "3de77cab-fde4-438d-956c-8381b6d003c2"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.509845 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "3de77cab-fde4-438d-956c-8381b6d003c2" (UID: "3de77cab-fde4-438d-956c-8381b6d003c2"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.522118 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3de77cab-fde4-438d-956c-8381b6d003c2" (UID: "3de77cab-fde4-438d-956c-8381b6d003c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.523548 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-config-data" (OuterVolumeSpecName: "config-data") pod "3de77cab-fde4-438d-956c-8381b6d003c2" (UID: "3de77cab-fde4-438d-956c-8381b6d003c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.596023 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-662gl\" (UniqueName: \"kubernetes.io/projected/3de77cab-fde4-438d-956c-8381b6d003c2-kube-api-access-662gl\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.596052 4578 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.596063 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.596072 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.596080 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.596089 4578 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3de77cab-fde4-438d-956c-8381b6d003c2-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.768878 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zmn6h" event={"ID":"3de77cab-fde4-438d-956c-8381b6d003c2","Type":"ContainerDied","Data":"ec09de033df5e0e07d1133685c9fe4da3d381f833d228b30fee0d3217d09a688"} Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.769155 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec09de033df5e0e07d1133685c9fe4da3d381f833d228b30fee0d3217d09a688" Oct 03 13:12:19 crc kubenswrapper[4578]: I1003 13:12:19.768971 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zmn6h" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.473624 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-zmn6h"] Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.479710 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-zmn6h"] Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.568324 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-jrdpv"] Oct 03 13:12:20 crc kubenswrapper[4578]: E1003 13:12:20.569203 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3de77cab-fde4-438d-956c-8381b6d003c2" containerName="keystone-bootstrap" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.569361 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="3de77cab-fde4-438d-956c-8381b6d003c2" containerName="keystone-bootstrap" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.569664 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="3de77cab-fde4-438d-956c-8381b6d003c2" containerName="keystone-bootstrap" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.571439 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.578707 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jrdpv"] Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.578950 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.579118 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.580382 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.580436 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wxw7h" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.729087 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-combined-ca-bundle\") pod \"keystone-bootstrap-jrdpv\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.729132 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-scripts\") pod \"keystone-bootstrap-jrdpv\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.729210 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-config-data\") pod \"keystone-bootstrap-jrdpv\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.729264 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-credential-keys\") pod \"keystone-bootstrap-jrdpv\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.729280 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-fernet-keys\") pod \"keystone-bootstrap-jrdpv\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.729334 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq528\" (UniqueName: \"kubernetes.io/projected/a9856a93-43ca-4f3d-96a5-50c85086ae91-kube-api-access-qq528\") pod \"keystone-bootstrap-jrdpv\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.831237 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq528\" (UniqueName: \"kubernetes.io/projected/a9856a93-43ca-4f3d-96a5-50c85086ae91-kube-api-access-qq528\") pod \"keystone-bootstrap-jrdpv\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.831334 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-combined-ca-bundle\") pod \"keystone-bootstrap-jrdpv\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.831356 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-scripts\") pod \"keystone-bootstrap-jrdpv\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.831404 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-config-data\") pod \"keystone-bootstrap-jrdpv\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.831436 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-credential-keys\") pod \"keystone-bootstrap-jrdpv\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.831451 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-fernet-keys\") pod \"keystone-bootstrap-jrdpv\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.836283 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-scripts\") pod \"keystone-bootstrap-jrdpv\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.844598 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-credential-keys\") pod \"keystone-bootstrap-jrdpv\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.845050 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-fernet-keys\") pod \"keystone-bootstrap-jrdpv\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.845076 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-config-data\") pod \"keystone-bootstrap-jrdpv\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.845655 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-combined-ca-bundle\") pod \"keystone-bootstrap-jrdpv\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.847596 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq528\" (UniqueName: \"kubernetes.io/projected/a9856a93-43ca-4f3d-96a5-50c85086ae91-kube-api-access-qq528\") pod \"keystone-bootstrap-jrdpv\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.903593 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:20 crc kubenswrapper[4578]: I1003 13:12:20.921122 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3de77cab-fde4-438d-956c-8381b6d003c2" path="/var/lib/kubelet/pods/3de77cab-fde4-438d-956c-8381b6d003c2/volumes" Oct 03 13:12:23 crc kubenswrapper[4578]: I1003 13:12:23.812307 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80","Type":"ContainerStarted","Data":"1e9a01f04842cca5dfcea8b57f22948fae9303fd28e03ead9afe067aaf096423"} Oct 03 13:12:27 crc kubenswrapper[4578]: I1003 13:12:27.180950 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" podUID="92cd4ef0-355f-4564-9590-6593431b678a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: i/o timeout" Oct 03 13:12:27 crc kubenswrapper[4578]: E1003 13:12:27.575334 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 03 13:12:27 crc kubenswrapper[4578]: E1003 13:12:27.575983 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n55dhb7h5b9h589h66dh54bh7h687hcbh667h54h67ch5b7h546h58hd5h68fh554h84hffh78h58h565h5c6h67dh85h5dbh74h5cdhcfhfdh5bq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w85j6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7fb966687c-gnkj8_openstack(30ecd9ab-e64e-49d3-bdf6-4f23af14afe6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:12:27 crc kubenswrapper[4578]: E1003 13:12:27.585502 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7fb966687c-gnkj8" podUID="30ecd9ab-e64e-49d3-bdf6-4f23af14afe6" Oct 03 13:12:27 crc kubenswrapper[4578]: E1003 13:12:27.969595 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Oct 03 13:12:27 crc kubenswrapper[4578]: E1003 13:12:27.969759 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n54h65ch577hc6h559h8fh5d9hb9h57h5dch85h5d6hd4hbch5dh67bhdch676h55dh548h598h5bdh87h5dchd7h5b9h59hd8h66h88hfhf7q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tbm4c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(6228c849-5824-4a21-8318-754c65d79a2e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.054917 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.174966 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-dns-svc\") pod \"92cd4ef0-355f-4564-9590-6593431b678a\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.175046 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-ovsdbserver-sb\") pod \"92cd4ef0-355f-4564-9590-6593431b678a\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.175093 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-dns-swift-storage-0\") pod \"92cd4ef0-355f-4564-9590-6593431b678a\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.175119 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-ovsdbserver-nb\") pod \"92cd4ef0-355f-4564-9590-6593431b678a\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.175225 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgl8s\" (UniqueName: \"kubernetes.io/projected/92cd4ef0-355f-4564-9590-6593431b678a-kube-api-access-tgl8s\") pod \"92cd4ef0-355f-4564-9590-6593431b678a\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.175250 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-config\") pod \"92cd4ef0-355f-4564-9590-6593431b678a\" (UID: \"92cd4ef0-355f-4564-9590-6593431b678a\") " Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.197431 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92cd4ef0-355f-4564-9590-6593431b678a-kube-api-access-tgl8s" (OuterVolumeSpecName: "kube-api-access-tgl8s") pod "92cd4ef0-355f-4564-9590-6593431b678a" (UID: "92cd4ef0-355f-4564-9590-6593431b678a"). InnerVolumeSpecName "kube-api-access-tgl8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.232880 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "92cd4ef0-355f-4564-9590-6593431b678a" (UID: "92cd4ef0-355f-4564-9590-6593431b678a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.234408 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "92cd4ef0-355f-4564-9590-6593431b678a" (UID: "92cd4ef0-355f-4564-9590-6593431b678a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.246833 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "92cd4ef0-355f-4564-9590-6593431b678a" (UID: "92cd4ef0-355f-4564-9590-6593431b678a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.247473 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "92cd4ef0-355f-4564-9590-6593431b678a" (UID: "92cd4ef0-355f-4564-9590-6593431b678a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.250844 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-config" (OuterVolumeSpecName: "config") pod "92cd4ef0-355f-4564-9590-6593431b678a" (UID: "92cd4ef0-355f-4564-9590-6593431b678a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.277324 4578 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.277443 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.277459 4578 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.277471 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.277481 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgl8s\" (UniqueName: \"kubernetes.io/projected/92cd4ef0-355f-4564-9590-6593431b678a-kube-api-access-tgl8s\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.277504 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92cd4ef0-355f-4564-9590-6593431b678a-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:28 crc kubenswrapper[4578]: E1003 13:12:28.645371 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Oct 03 13:12:28 crc kubenswrapper[4578]: E1003 13:12:28.645526 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j9lsz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-j9scz_openstack(eacf42c3-8e6f-4304-b17b-de32359e8d76): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:12:28 crc kubenswrapper[4578]: E1003 13:12:28.646717 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-j9scz" podUID="eacf42c3-8e6f-4304-b17b-de32359e8d76" Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.858944 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.859847 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" event={"ID":"92cd4ef0-355f-4564-9590-6593431b678a","Type":"ContainerDied","Data":"8ae285be63c4d9907281b7363ed1fd92f0207006be3f49d57f867a9f236c9400"} Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.859893 4578 scope.go:117] "RemoveContainer" containerID="67a6dda3e987feb1381d7012ff7fd320626f3627c9b84978669afe0518d9e937" Oct 03 13:12:28 crc kubenswrapper[4578]: E1003 13:12:28.860605 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-j9scz" podUID="eacf42c3-8e6f-4304-b17b-de32359e8d76" Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.899187 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-s7mxj"] Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.905711 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-s7mxj"] Oct 03 13:12:28 crc kubenswrapper[4578]: I1003 13:12:28.921007 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92cd4ef0-355f-4564-9590-6593431b678a" path="/var/lib/kubelet/pods/92cd4ef0-355f-4564-9590-6593431b678a/volumes" Oct 03 13:12:31 crc kubenswrapper[4578]: I1003 13:12:31.891505 4578 generic.go:334] "Generic (PLEG): container finished" podID="0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924" containerID="0abf83735c6600eb12b2aad4c07c5128d9e4fed23e86e7fec2aadf575ac91bcb" exitCode=0 Oct 03 13:12:31 crc kubenswrapper[4578]: I1003 13:12:31.892329 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-85psh" event={"ID":"0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924","Type":"ContainerDied","Data":"0abf83735c6600eb12b2aad4c07c5128d9e4fed23e86e7fec2aadf575ac91bcb"} Oct 03 13:12:32 crc kubenswrapper[4578]: I1003 13:12:32.181665 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-s7mxj" podUID="92cd4ef0-355f-4564-9590-6593431b678a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: i/o timeout" Oct 03 13:12:35 crc kubenswrapper[4578]: I1003 13:12:35.091740 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:12:35 crc kubenswrapper[4578]: I1003 13:12:35.092282 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:12:35 crc kubenswrapper[4578]: I1003 13:12:35.092321 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 13:12:35 crc kubenswrapper[4578]: I1003 13:12:35.093100 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ad105f3cc883a1c5d61b8c9107a9b7139beb580cccd8fece952ff237bc9b856e"} pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:12:35 crc kubenswrapper[4578]: I1003 13:12:35.093155 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" containerID="cri-o://ad105f3cc883a1c5d61b8c9107a9b7139beb580cccd8fece952ff237bc9b856e" gracePeriod=600 Oct 03 13:12:36 crc kubenswrapper[4578]: I1003 13:12:36.939723 4578 generic.go:334] "Generic (PLEG): container finished" podID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerID="ad105f3cc883a1c5d61b8c9107a9b7139beb580cccd8fece952ff237bc9b856e" exitCode=0 Oct 03 13:12:36 crc kubenswrapper[4578]: I1003 13:12:36.939800 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerDied","Data":"ad105f3cc883a1c5d61b8c9107a9b7139beb580cccd8fece952ff237bc9b856e"} Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.550493 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fb966687c-gnkj8" Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.564101 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-85psh" Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.601761 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-scripts\") pod \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\" (UID: \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\") " Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.601847 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-logs\") pod \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\" (UID: \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\") " Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.601934 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-config-data\") pod \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\" (UID: \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\") " Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.601987 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924-combined-ca-bundle\") pod \"0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924\" (UID: \"0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924\") " Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.602047 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924-config\") pod \"0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924\" (UID: \"0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924\") " Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.602094 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-horizon-secret-key\") pod \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\" (UID: \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\") " Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.602146 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7zk9\" (UniqueName: \"kubernetes.io/projected/0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924-kube-api-access-d7zk9\") pod \"0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924\" (UID: \"0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924\") " Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.602187 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w85j6\" (UniqueName: \"kubernetes.io/projected/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-kube-api-access-w85j6\") pod \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\" (UID: \"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6\") " Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.619708 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-scripts" (OuterVolumeSpecName: "scripts") pod "30ecd9ab-e64e-49d3-bdf6-4f23af14afe6" (UID: "30ecd9ab-e64e-49d3-bdf6-4f23af14afe6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.623477 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-config-data" (OuterVolumeSpecName: "config-data") pod "30ecd9ab-e64e-49d3-bdf6-4f23af14afe6" (UID: "30ecd9ab-e64e-49d3-bdf6-4f23af14afe6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.623670 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-kube-api-access-w85j6" (OuterVolumeSpecName: "kube-api-access-w85j6") pod "30ecd9ab-e64e-49d3-bdf6-4f23af14afe6" (UID: "30ecd9ab-e64e-49d3-bdf6-4f23af14afe6"). InnerVolumeSpecName "kube-api-access-w85j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.625017 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-logs" (OuterVolumeSpecName: "logs") pod "30ecd9ab-e64e-49d3-bdf6-4f23af14afe6" (UID: "30ecd9ab-e64e-49d3-bdf6-4f23af14afe6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.628092 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "30ecd9ab-e64e-49d3-bdf6-4f23af14afe6" (UID: "30ecd9ab-e64e-49d3-bdf6-4f23af14afe6"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.635519 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924-kube-api-access-d7zk9" (OuterVolumeSpecName: "kube-api-access-d7zk9") pod "0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924" (UID: "0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924"). InnerVolumeSpecName "kube-api-access-d7zk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.655172 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924-config" (OuterVolumeSpecName: "config") pod "0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924" (UID: "0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.658103 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924" (UID: "0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.704255 4578 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.704303 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.704316 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.704328 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.704338 4578 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.704347 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7zk9\" (UniqueName: \"kubernetes.io/projected/0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924-kube-api-access-d7zk9\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.704360 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w85j6\" (UniqueName: \"kubernetes.io/projected/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-kube-api-access-w85j6\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.704370 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.965598 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fb966687c-gnkj8" event={"ID":"30ecd9ab-e64e-49d3-bdf6-4f23af14afe6","Type":"ContainerDied","Data":"0b8ca28b2f9ee9e2d75784f45e79422bff81fec27b015a1eaf128c326e8f2a82"} Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.965714 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fb966687c-gnkj8" Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.969972 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-85psh" event={"ID":"0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924","Type":"ContainerDied","Data":"f0b8a13c8daf999d9420867555f31332de21fc270e85dae2adefe24914991d27"} Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.970021 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0b8a13c8daf999d9420867555f31332de21fc270e85dae2adefe24914991d27" Oct 03 13:12:39 crc kubenswrapper[4578]: I1003 13:12:39.970091 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-85psh" Oct 03 13:12:40 crc kubenswrapper[4578]: I1003 13:12:40.048598 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7fb966687c-gnkj8"] Oct 03 13:12:40 crc kubenswrapper[4578]: I1003 13:12:40.057331 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7fb966687c-gnkj8"] Oct 03 13:12:40 crc kubenswrapper[4578]: I1003 13:12:40.725170 4578 scope.go:117] "RemoveContainer" containerID="67712ace0dca9eebde84c2e8eef14d529293e75e8b113920ab2c4f941ca4c5d5" Oct 03 13:12:40 crc kubenswrapper[4578]: I1003 13:12:40.876588 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-phtfx"] Oct 03 13:12:40 crc kubenswrapper[4578]: E1003 13:12:40.877440 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 03 13:12:40 crc kubenswrapper[4578]: E1003 13:12:40.877574 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-55whk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-q6ppj_openstack(6b310034-f02f-48ac-a47a-4f41f15379a5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 03 13:12:40 crc kubenswrapper[4578]: E1003 13:12:40.883176 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-q6ppj" podUID="6b310034-f02f-48ac-a47a-4f41f15379a5" Oct 03 13:12:40 crc kubenswrapper[4578]: E1003 13:12:40.883903 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924" containerName="neutron-db-sync" Oct 03 13:12:40 crc kubenswrapper[4578]: I1003 13:12:40.883920 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924" containerName="neutron-db-sync" Oct 03 13:12:40 crc kubenswrapper[4578]: E1003 13:12:40.883940 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92cd4ef0-355f-4564-9590-6593431b678a" containerName="init" Oct 03 13:12:40 crc kubenswrapper[4578]: I1003 13:12:40.883954 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="92cd4ef0-355f-4564-9590-6593431b678a" containerName="init" Oct 03 13:12:40 crc kubenswrapper[4578]: E1003 13:12:40.883973 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92cd4ef0-355f-4564-9590-6593431b678a" containerName="dnsmasq-dns" Oct 03 13:12:40 crc kubenswrapper[4578]: I1003 13:12:40.883979 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="92cd4ef0-355f-4564-9590-6593431b678a" containerName="dnsmasq-dns" Oct 03 13:12:40 crc kubenswrapper[4578]: I1003 13:12:40.884140 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="92cd4ef0-355f-4564-9590-6593431b678a" containerName="dnsmasq-dns" Oct 03 13:12:40 crc kubenswrapper[4578]: I1003 13:12:40.884158 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924" containerName="neutron-db-sync" Oct 03 13:12:40 crc kubenswrapper[4578]: I1003 13:12:40.887128 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:40 crc kubenswrapper[4578]: I1003 13:12:40.935266 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-phtfx\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:40 crc kubenswrapper[4578]: I1003 13:12:40.935603 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-phtfx\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:40 crc kubenswrapper[4578]: I1003 13:12:40.935652 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnqwm\" (UniqueName: \"kubernetes.io/projected/b85d10a6-888d-4862-986f-5e44b8d3d229-kube-api-access-fnqwm\") pod \"dnsmasq-dns-84b966f6c9-phtfx\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:40 crc kubenswrapper[4578]: I1003 13:12:40.935686 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-phtfx\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:40 crc kubenswrapper[4578]: I1003 13:12:40.935720 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-phtfx\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:40 crc kubenswrapper[4578]: I1003 13:12:40.935785 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-config\") pod \"dnsmasq-dns-84b966f6c9-phtfx\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:40 crc kubenswrapper[4578]: I1003 13:12:40.965843 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30ecd9ab-e64e-49d3-bdf6-4f23af14afe6" path="/var/lib/kubelet/pods/30ecd9ab-e64e-49d3-bdf6-4f23af14afe6/volumes" Oct 03 13:12:40 crc kubenswrapper[4578]: I1003 13:12:40.966241 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-phtfx"] Oct 03 13:12:41 crc kubenswrapper[4578]: E1003 13:12:41.010704 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-q6ppj" podUID="6b310034-f02f-48ac-a47a-4f41f15379a5" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.036809 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-config\") pod \"dnsmasq-dns-84b966f6c9-phtfx\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.036869 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-phtfx\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.036907 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-phtfx\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.036946 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnqwm\" (UniqueName: \"kubernetes.io/projected/b85d10a6-888d-4862-986f-5e44b8d3d229-kube-api-access-fnqwm\") pod \"dnsmasq-dns-84b966f6c9-phtfx\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.036979 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-phtfx\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.037020 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-phtfx\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.038901 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-phtfx\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.038932 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-config\") pod \"dnsmasq-dns-84b966f6c9-phtfx\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.039598 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-phtfx\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.039660 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-phtfx\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.040107 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-phtfx\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.092485 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnqwm\" (UniqueName: \"kubernetes.io/projected/b85d10a6-888d-4862-986f-5e44b8d3d229-kube-api-access-fnqwm\") pod \"dnsmasq-dns-84b966f6c9-phtfx\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.112902 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6d67444cbd-9qgkn"] Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.115058 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.118424 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.118650 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.118846 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.118848 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-wbbsn" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.126221 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6d67444cbd-9qgkn"] Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.140162 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-config\") pod \"neutron-6d67444cbd-9qgkn\" (UID: \"101089c9-6f91-42fb-a5ab-d907c80ff069\") " pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.140241 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2n5t\" (UniqueName: \"kubernetes.io/projected/101089c9-6f91-42fb-a5ab-d907c80ff069-kube-api-access-s2n5t\") pod \"neutron-6d67444cbd-9qgkn\" (UID: \"101089c9-6f91-42fb-a5ab-d907c80ff069\") " pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.140259 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-ovndb-tls-certs\") pod \"neutron-6d67444cbd-9qgkn\" (UID: \"101089c9-6f91-42fb-a5ab-d907c80ff069\") " pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.140337 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-httpd-config\") pod \"neutron-6d67444cbd-9qgkn\" (UID: \"101089c9-6f91-42fb-a5ab-d907c80ff069\") " pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.140363 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-combined-ca-bundle\") pod \"neutron-6d67444cbd-9qgkn\" (UID: \"101089c9-6f91-42fb-a5ab-d907c80ff069\") " pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.237222 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.241951 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-httpd-config\") pod \"neutron-6d67444cbd-9qgkn\" (UID: \"101089c9-6f91-42fb-a5ab-d907c80ff069\") " pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.242003 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-combined-ca-bundle\") pod \"neutron-6d67444cbd-9qgkn\" (UID: \"101089c9-6f91-42fb-a5ab-d907c80ff069\") " pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.242058 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-config\") pod \"neutron-6d67444cbd-9qgkn\" (UID: \"101089c9-6f91-42fb-a5ab-d907c80ff069\") " pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.242109 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2n5t\" (UniqueName: \"kubernetes.io/projected/101089c9-6f91-42fb-a5ab-d907c80ff069-kube-api-access-s2n5t\") pod \"neutron-6d67444cbd-9qgkn\" (UID: \"101089c9-6f91-42fb-a5ab-d907c80ff069\") " pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.242126 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-ovndb-tls-certs\") pod \"neutron-6d67444cbd-9qgkn\" (UID: \"101089c9-6f91-42fb-a5ab-d907c80ff069\") " pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.250396 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-combined-ca-bundle\") pod \"neutron-6d67444cbd-9qgkn\" (UID: \"101089c9-6f91-42fb-a5ab-d907c80ff069\") " pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.250570 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-config\") pod \"neutron-6d67444cbd-9qgkn\" (UID: \"101089c9-6f91-42fb-a5ab-d907c80ff069\") " pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.266419 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-ovndb-tls-certs\") pod \"neutron-6d67444cbd-9qgkn\" (UID: \"101089c9-6f91-42fb-a5ab-d907c80ff069\") " pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.269199 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2n5t\" (UniqueName: \"kubernetes.io/projected/101089c9-6f91-42fb-a5ab-d907c80ff069-kube-api-access-s2n5t\") pod \"neutron-6d67444cbd-9qgkn\" (UID: \"101089c9-6f91-42fb-a5ab-d907c80ff069\") " pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.273431 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-httpd-config\") pod \"neutron-6d67444cbd-9qgkn\" (UID: \"101089c9-6f91-42fb-a5ab-d907c80ff069\") " pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.415581 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68684899bd-2cwll"] Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.450567 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.571249 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:12:41 crc kubenswrapper[4578]: W1003 13:12:41.764284 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd6a1a68_a60c_41d3_a0e9_ca51feef6df5.slice/crio-be1412f89b62042a047caf0003a928e2c4fc004839bcc7cf4efc9425bc025dc8 WatchSource:0}: Error finding container be1412f89b62042a047caf0003a928e2c4fc004839bcc7cf4efc9425bc025dc8: Status 404 returned error can't find the container with id be1412f89b62042a047caf0003a928e2c4fc004839bcc7cf4efc9425bc025dc8 Oct 03 13:12:41 crc kubenswrapper[4578]: I1003 13:12:41.860336 4578 scope.go:117] "RemoveContainer" containerID="415155e4360685a3545a1494ed131d83906f75047c2e61272798309c3d089d8b" Oct 03 13:12:42 crc kubenswrapper[4578]: I1003 13:12:42.045658 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68684899bd-2cwll" event={"ID":"d49d623e-2cd6-4786-b1da-34b78b89a134","Type":"ContainerStarted","Data":"165603208d8d99781a2adf6b2fe48916cdc84c977ff4e0db61f1f63c91f6e4df"} Oct 03 13:12:42 crc kubenswrapper[4578]: I1003 13:12:42.067404 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5","Type":"ContainerStarted","Data":"be1412f89b62042a047caf0003a928e2c4fc004839bcc7cf4efc9425bc025dc8"} Oct 03 13:12:42 crc kubenswrapper[4578]: I1003 13:12:42.335015 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jrdpv"] Oct 03 13:12:42 crc kubenswrapper[4578]: I1003 13:12:42.388753 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7b58cf9568-f8nng"] Oct 03 13:12:42 crc kubenswrapper[4578]: I1003 13:12:42.704411 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-phtfx"] Oct 03 13:12:42 crc kubenswrapper[4578]: W1003 13:12:42.729088 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb85d10a6_888d_4862_986f_5e44b8d3d229.slice/crio-5ef03001cc424673a570bab42eb85076c5d758abfe4c48e50a9378476d665592 WatchSource:0}: Error finding container 5ef03001cc424673a570bab42eb85076c5d758abfe4c48e50a9378476d665592: Status 404 returned error can't find the container with id 5ef03001cc424673a570bab42eb85076c5d758abfe4c48e50a9378476d665592 Oct 03 13:12:42 crc kubenswrapper[4578]: I1003 13:12:42.825875 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6d67444cbd-9qgkn"] Oct 03 13:12:43 crc kubenswrapper[4578]: I1003 13:12:43.112133 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" event={"ID":"b85d10a6-888d-4862-986f-5e44b8d3d229","Type":"ContainerStarted","Data":"5ef03001cc424673a570bab42eb85076c5d758abfe4c48e50a9378476d665592"} Oct 03 13:12:43 crc kubenswrapper[4578]: I1003 13:12:43.138837 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6d67444cbd-9qgkn" event={"ID":"101089c9-6f91-42fb-a5ab-d907c80ff069","Type":"ContainerStarted","Data":"bbc5a55643316bbce08815fc7089ca7afc837fdc9726a9b22385bbe451a8e6fe"} Oct 03 13:12:43 crc kubenswrapper[4578]: I1003 13:12:43.152404 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6228c849-5824-4a21-8318-754c65d79a2e","Type":"ContainerStarted","Data":"b45d39e61e8fc0942f781f361e7f58c723585ade153a008212d11232868eaf4d"} Oct 03 13:12:43 crc kubenswrapper[4578]: I1003 13:12:43.167847 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jrdpv" event={"ID":"a9856a93-43ca-4f3d-96a5-50c85086ae91","Type":"ContainerStarted","Data":"4c31f2c7bec5ec73b4641d8ab5761d2f40e161d8db396a1f5ffe89700025a8c9"} Oct 03 13:12:43 crc kubenswrapper[4578]: I1003 13:12:43.167894 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jrdpv" event={"ID":"a9856a93-43ca-4f3d-96a5-50c85086ae91","Type":"ContainerStarted","Data":"e6518571c6e27ae02edc2365e1598cfd78d3140aca53fac6281ace75cb6eb008"} Oct 03 13:12:43 crc kubenswrapper[4578]: I1003 13:12:43.184525 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c684ff4d9-2h9cq" podUID="1b0f9e3d-a619-437f-9339-d61b6ffb8d4b" containerName="horizon-log" containerID="cri-o://5402c4c9ddc18e14fe296ec7864a7e4a97e4da6535530acc67d68775074d1046" gracePeriod=30 Oct 03 13:12:43 crc kubenswrapper[4578]: I1003 13:12:43.184710 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c684ff4d9-2h9cq" event={"ID":"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b","Type":"ContainerStarted","Data":"80929f7ff9f2db6421b73861631c698d81ca2c15a2ea97522dc9d1eb510196b2"} Oct 03 13:12:43 crc kubenswrapper[4578]: I1003 13:12:43.184989 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c684ff4d9-2h9cq" event={"ID":"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b","Type":"ContainerStarted","Data":"5402c4c9ddc18e14fe296ec7864a7e4a97e4da6535530acc67d68775074d1046"} Oct 03 13:12:43 crc kubenswrapper[4578]: I1003 13:12:43.184832 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c684ff4d9-2h9cq" podUID="1b0f9e3d-a619-437f-9339-d61b6ffb8d4b" containerName="horizon" containerID="cri-o://80929f7ff9f2db6421b73861631c698d81ca2c15a2ea97522dc9d1eb510196b2" gracePeriod=30 Oct 03 13:12:43 crc kubenswrapper[4578]: I1003 13:12:43.190013 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-jrdpv" podStartSLOduration=23.189993173 podStartE2EDuration="23.189993173s" podCreationTimestamp="2025-10-03 13:12:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:12:43.185417061 +0000 UTC m=+1298.983889255" watchObservedRunningTime="2025-10-03 13:12:43.189993173 +0000 UTC m=+1298.988465357" Oct 03 13:12:43 crc kubenswrapper[4578]: I1003 13:12:43.217676 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95"} Oct 03 13:12:43 crc kubenswrapper[4578]: I1003 13:12:43.235183 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7c684ff4d9-2h9cq" podStartSLOduration=5.0012989 podStartE2EDuration="37.235149261s" podCreationTimestamp="2025-10-03 13:12:06 +0000 UTC" firstStartedPulling="2025-10-03 13:12:08.4687651 +0000 UTC m=+1264.267237284" lastFinishedPulling="2025-10-03 13:12:40.702615461 +0000 UTC m=+1296.501087645" observedRunningTime="2025-10-03 13:12:43.209049027 +0000 UTC m=+1299.007521211" watchObservedRunningTime="2025-10-03 13:12:43.235149261 +0000 UTC m=+1299.033621445" Oct 03 13:12:43 crc kubenswrapper[4578]: I1003 13:12:43.245651 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gv9w6" event={"ID":"eb6557a6-8665-4a81-afec-4ccaef43b801","Type":"ContainerStarted","Data":"d3deb53df311f3445d8b678731a0915eeb05d092fbb84ef502995999cad84baf"} Oct 03 13:12:43 crc kubenswrapper[4578]: I1003 13:12:43.254731 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68684899bd-2cwll" event={"ID":"d49d623e-2cd6-4786-b1da-34b78b89a134","Type":"ContainerStarted","Data":"8cad347bd426ae4fd15883ee421e528b250579007674d8ddbdb55030c404b792"} Oct 03 13:12:43 crc kubenswrapper[4578]: I1003 13:12:43.271614 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b58cf9568-f8nng" event={"ID":"a54aea6a-c051-4832-b764-3ef53a9cc3a1","Type":"ContainerStarted","Data":"de051c4c2fdbf56aef0e53a639dcd8953e6218bb9bdef3f691e52909ccf616ec"} Oct 03 13:12:43 crc kubenswrapper[4578]: I1003 13:12:43.271690 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b58cf9568-f8nng" event={"ID":"a54aea6a-c051-4832-b764-3ef53a9cc3a1","Type":"ContainerStarted","Data":"e283c28782e5f44345615ef81d7b97769693cfdd7bce448773683ccbd06e5f51"} Oct 03 13:12:43 crc kubenswrapper[4578]: I1003 13:12:43.301332 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d5f486769-rxsds" event={"ID":"a567645b-5ae1-46a0-86b1-e24b1592e060","Type":"ContainerStarted","Data":"00d4bed9ee5fe90eebe7ce2ec9be50c6902436a20863280d15aafd3f0158f87d"} Oct 03 13:12:43 crc kubenswrapper[4578]: I1003 13:12:43.338461 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-gv9w6" podStartSLOduration=7.008216797 podStartE2EDuration="37.338433989s" podCreationTimestamp="2025-10-03 13:12:06 +0000 UTC" firstStartedPulling="2025-10-03 13:12:09.123331284 +0000 UTC m=+1264.921803469" lastFinishedPulling="2025-10-03 13:12:39.453548477 +0000 UTC m=+1295.252020661" observedRunningTime="2025-10-03 13:12:43.317945441 +0000 UTC m=+1299.116417635" watchObservedRunningTime="2025-10-03 13:12:43.338433989 +0000 UTC m=+1299.136906173" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.344858 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6d67444cbd-9qgkn" event={"ID":"101089c9-6f91-42fb-a5ab-d907c80ff069","Type":"ContainerStarted","Data":"e73cf2c70f9bc072fedc51586fa3a7d7b4c59719b48f280707a595fd70f5d102"} Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.369773 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d5f486769-rxsds" event={"ID":"a567645b-5ae1-46a0-86b1-e24b1592e060","Type":"ContainerStarted","Data":"4b3f4f06273a71501b5530c10beab91d989254af38cef506f55056b9e26f0b06"} Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.369992 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-d5f486769-rxsds" podUID="a567645b-5ae1-46a0-86b1-e24b1592e060" containerName="horizon-log" containerID="cri-o://00d4bed9ee5fe90eebe7ce2ec9be50c6902436a20863280d15aafd3f0158f87d" gracePeriod=30 Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.370067 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-d5f486769-rxsds" podUID="a567645b-5ae1-46a0-86b1-e24b1592e060" containerName="horizon" containerID="cri-o://4b3f4f06273a71501b5530c10beab91d989254af38cef506f55056b9e26f0b06" gracePeriod=30 Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.392233 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5","Type":"ContainerStarted","Data":"c4550bc1dabcb2df378a1554b302e86b33ee5bc061cd46dc7496c1d0887a7210"} Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.404491 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-d5f486769-rxsds" podStartSLOduration=4.018079456 podStartE2EDuration="34.404474489s" podCreationTimestamp="2025-10-03 13:12:10 +0000 UTC" firstStartedPulling="2025-10-03 13:12:11.374551318 +0000 UTC m=+1267.173023502" lastFinishedPulling="2025-10-03 13:12:41.760946351 +0000 UTC m=+1297.559418535" observedRunningTime="2025-10-03 13:12:44.401276419 +0000 UTC m=+1300.199748603" watchObservedRunningTime="2025-10-03 13:12:44.404474489 +0000 UTC m=+1300.202946683" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.405302 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68684899bd-2cwll" event={"ID":"d49d623e-2cd6-4786-b1da-34b78b89a134","Type":"ContainerStarted","Data":"043ba4870e4db2272214244e538587f588862b6e71b023c4ea7dfa6638da57bd"} Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.414524 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b58cf9568-f8nng" event={"ID":"a54aea6a-c051-4832-b764-3ef53a9cc3a1","Type":"ContainerStarted","Data":"033a127d7b3dc6a5469b25fc2682629ad13d3e1ed8dd904aad8edc6112ea940b"} Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.433774 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80","Type":"ContainerStarted","Data":"b87bf43c564aa0714f110549b69b1d0bf12d6c2035f2dcecb4144e9b98bf2b9b"} Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.440702 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-68684899bd-2cwll" podStartSLOduration=28.440680398 podStartE2EDuration="28.440680398s" podCreationTimestamp="2025-10-03 13:12:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:12:44.434883107 +0000 UTC m=+1300.233355301" watchObservedRunningTime="2025-10-03 13:12:44.440680398 +0000 UTC m=+1300.239152582" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.466422 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7b58cf9568-f8nng" podStartSLOduration=28.46640636 podStartE2EDuration="28.46640636s" podCreationTimestamp="2025-10-03 13:12:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:12:44.45998158 +0000 UTC m=+1300.258453764" watchObservedRunningTime="2025-10-03 13:12:44.46640636 +0000 UTC m=+1300.264878544" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.616098 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-d7d5c9485-l2ww5"] Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.618266 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.621989 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.622173 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.637245 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d7d5c9485-l2ww5"] Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.689907 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2711a2e9-8d7a-4942-b762-f211d519e0f4-public-tls-certs\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.689960 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w84lh\" (UniqueName: \"kubernetes.io/projected/2711a2e9-8d7a-4942-b762-f211d519e0f4-kube-api-access-w84lh\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.689984 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2711a2e9-8d7a-4942-b762-f211d519e0f4-combined-ca-bundle\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.690004 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2711a2e9-8d7a-4942-b762-f211d519e0f4-httpd-config\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.690040 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2711a2e9-8d7a-4942-b762-f211d519e0f4-internal-tls-certs\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.690070 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2711a2e9-8d7a-4942-b762-f211d519e0f4-ovndb-tls-certs\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.690103 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2711a2e9-8d7a-4942-b762-f211d519e0f4-config\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.791399 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2711a2e9-8d7a-4942-b762-f211d519e0f4-ovndb-tls-certs\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.791492 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2711a2e9-8d7a-4942-b762-f211d519e0f4-config\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.791623 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2711a2e9-8d7a-4942-b762-f211d519e0f4-public-tls-certs\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.791671 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w84lh\" (UniqueName: \"kubernetes.io/projected/2711a2e9-8d7a-4942-b762-f211d519e0f4-kube-api-access-w84lh\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.791700 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2711a2e9-8d7a-4942-b762-f211d519e0f4-combined-ca-bundle\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.791724 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2711a2e9-8d7a-4942-b762-f211d519e0f4-httpd-config\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.791774 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2711a2e9-8d7a-4942-b762-f211d519e0f4-internal-tls-certs\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.798331 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2711a2e9-8d7a-4942-b762-f211d519e0f4-ovndb-tls-certs\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.799127 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2711a2e9-8d7a-4942-b762-f211d519e0f4-internal-tls-certs\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.800368 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2711a2e9-8d7a-4942-b762-f211d519e0f4-combined-ca-bundle\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.803020 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2711a2e9-8d7a-4942-b762-f211d519e0f4-public-tls-certs\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.812778 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2711a2e9-8d7a-4942-b762-f211d519e0f4-config\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.817499 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w84lh\" (UniqueName: \"kubernetes.io/projected/2711a2e9-8d7a-4942-b762-f211d519e0f4-kube-api-access-w84lh\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:44 crc kubenswrapper[4578]: I1003 13:12:44.824439 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2711a2e9-8d7a-4942-b762-f211d519e0f4-httpd-config\") pod \"neutron-d7d5c9485-l2ww5\" (UID: \"2711a2e9-8d7a-4942-b762-f211d519e0f4\") " pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:45 crc kubenswrapper[4578]: I1003 13:12:45.051209 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:45 crc kubenswrapper[4578]: I1003 13:12:45.455944 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80","Type":"ContainerStarted","Data":"313818f2faab0b600157582f66168ad9ab8c897bda250d026f7b4356128816d1"} Oct 03 13:12:45 crc kubenswrapper[4578]: I1003 13:12:45.456240 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80" containerName="glance-log" containerID="cri-o://b87bf43c564aa0714f110549b69b1d0bf12d6c2035f2dcecb4144e9b98bf2b9b" gracePeriod=30 Oct 03 13:12:45 crc kubenswrapper[4578]: I1003 13:12:45.456506 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80" containerName="glance-httpd" containerID="cri-o://313818f2faab0b600157582f66168ad9ab8c897bda250d026f7b4356128816d1" gracePeriod=30 Oct 03 13:12:45 crc kubenswrapper[4578]: I1003 13:12:45.459976 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6d67444cbd-9qgkn" event={"ID":"101089c9-6f91-42fb-a5ab-d907c80ff069","Type":"ContainerStarted","Data":"6617906f8965d5b3d283bc14710803835bf60a5cb3e056a9c10f0e336b4cdff7"} Oct 03 13:12:45 crc kubenswrapper[4578]: I1003 13:12:45.460670 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:12:45 crc kubenswrapper[4578]: I1003 13:12:45.474867 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-j9scz" event={"ID":"eacf42c3-8e6f-4304-b17b-de32359e8d76","Type":"ContainerStarted","Data":"5fa1c518576e9d2ffc9dda2b0ef5adaf2cf39618bda67a1eee0fb45f6bd0f0f4"} Oct 03 13:12:45 crc kubenswrapper[4578]: I1003 13:12:45.481492 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=31.481468931 podStartE2EDuration="31.481468931s" podCreationTimestamp="2025-10-03 13:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:12:45.475223046 +0000 UTC m=+1301.273695230" watchObservedRunningTime="2025-10-03 13:12:45.481468931 +0000 UTC m=+1301.279941115" Oct 03 13:12:45 crc kubenswrapper[4578]: I1003 13:12:45.485404 4578 generic.go:334] "Generic (PLEG): container finished" podID="b85d10a6-888d-4862-986f-5e44b8d3d229" containerID="90e2e3f7ccedd25cf461c8c8da59d658d9b7cc0793e9947c3ceafc7418d8ad90" exitCode=0 Oct 03 13:12:45 crc kubenswrapper[4578]: I1003 13:12:45.486929 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" event={"ID":"b85d10a6-888d-4862-986f-5e44b8d3d229","Type":"ContainerDied","Data":"90e2e3f7ccedd25cf461c8c8da59d658d9b7cc0793e9947c3ceafc7418d8ad90"} Oct 03 13:12:45 crc kubenswrapper[4578]: I1003 13:12:45.508783 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6d67444cbd-9qgkn" podStartSLOduration=4.508766302 podStartE2EDuration="4.508766302s" podCreationTimestamp="2025-10-03 13:12:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:12:45.49427872 +0000 UTC m=+1301.292750904" watchObservedRunningTime="2025-10-03 13:12:45.508766302 +0000 UTC m=+1301.307238486" Oct 03 13:12:45 crc kubenswrapper[4578]: I1003 13:12:45.529551 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-j9scz" podStartSLOduration=4.54647945 podStartE2EDuration="39.529531999s" podCreationTimestamp="2025-10-03 13:12:06 +0000 UTC" firstStartedPulling="2025-10-03 13:12:08.639390041 +0000 UTC m=+1264.437862225" lastFinishedPulling="2025-10-03 13:12:43.62244259 +0000 UTC m=+1299.420914774" observedRunningTime="2025-10-03 13:12:45.518445373 +0000 UTC m=+1301.316917567" watchObservedRunningTime="2025-10-03 13:12:45.529531999 +0000 UTC m=+1301.328004183" Oct 03 13:12:45 crc kubenswrapper[4578]: I1003 13:12:45.762334 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d7d5c9485-l2ww5"] Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.186428 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.229589 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.235999 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-public-tls-certs\") pod \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.236076 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-config-data\") pod \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.236114 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-logs\") pod \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.236204 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-httpd-run\") pod \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.236228 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfssl\" (UniqueName: \"kubernetes.io/projected/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-kube-api-access-lfssl\") pod \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.236277 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-scripts\") pod \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.236327 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-combined-ca-bundle\") pod \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\" (UID: \"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80\") " Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.242446 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-logs" (OuterVolumeSpecName: "logs") pod "3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80" (UID: "3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.244718 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80" (UID: "3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.267963 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-kube-api-access-lfssl" (OuterVolumeSpecName: "kube-api-access-lfssl") pod "3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80" (UID: "3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80"). InnerVolumeSpecName "kube-api-access-lfssl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.268059 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80" (UID: "3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.276549 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-scripts" (OuterVolumeSpecName: "scripts") pod "3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80" (UID: "3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.339661 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfssl\" (UniqueName: \"kubernetes.io/projected/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-kube-api-access-lfssl\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.339705 4578 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.339719 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.339754 4578 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.339766 4578 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.407433 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80" (UID: "3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.444756 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.458383 4578 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.498352 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-config-data" (OuterVolumeSpecName: "config-data") pod "3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80" (UID: "3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.512108 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" event={"ID":"b85d10a6-888d-4862-986f-5e44b8d3d229","Type":"ContainerStarted","Data":"460acd0e4b7cdbdda5a53c34a461e8bf587fbb22d1ee478ae0ec8f7d216a1abc"} Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.514043 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.536075 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7c684ff4d9-2h9cq" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.540436 4578 generic.go:334] "Generic (PLEG): container finished" podID="3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80" containerID="313818f2faab0b600157582f66168ad9ab8c897bda250d026f7b4356128816d1" exitCode=143 Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.540476 4578 generic.go:334] "Generic (PLEG): container finished" podID="3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80" containerID="b87bf43c564aa0714f110549b69b1d0bf12d6c2035f2dcecb4144e9b98bf2b9b" exitCode=143 Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.540514 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80","Type":"ContainerDied","Data":"313818f2faab0b600157582f66168ad9ab8c897bda250d026f7b4356128816d1"} Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.540557 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80","Type":"ContainerDied","Data":"b87bf43c564aa0714f110549b69b1d0bf12d6c2035f2dcecb4144e9b98bf2b9b"} Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.540568 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80","Type":"ContainerDied","Data":"1e9a01f04842cca5dfcea8b57f22948fae9303fd28e03ead9afe067aaf096423"} Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.540584 4578 scope.go:117] "RemoveContainer" containerID="313818f2faab0b600157582f66168ad9ab8c897bda250d026f7b4356128816d1" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.540756 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.544724 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80" (UID: "3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.545594 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" podStartSLOduration=6.545578541 podStartE2EDuration="6.545578541s" podCreationTimestamp="2025-10-03 13:12:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:12:46.540684858 +0000 UTC m=+1302.339157052" watchObservedRunningTime="2025-10-03 13:12:46.545578541 +0000 UTC m=+1302.344050725" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.545953 4578 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.545977 4578 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.545987 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.552679 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5","Type":"ContainerStarted","Data":"4204fc6b484771af3e2d7cbbbd547b7f401dba13ac6e8bcf7747e91bc1d2b151"} Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.552822 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="fd6a1a68-a60c-41d3-a0e9-ca51feef6df5" containerName="glance-log" containerID="cri-o://c4550bc1dabcb2df378a1554b302e86b33ee5bc061cd46dc7496c1d0887a7210" gracePeriod=30 Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.552908 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="fd6a1a68-a60c-41d3-a0e9-ca51feef6df5" containerName="glance-httpd" containerID="cri-o://4204fc6b484771af3e2d7cbbbd547b7f401dba13ac6e8bcf7747e91bc1d2b151" gracePeriod=30 Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.574235 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d7d5c9485-l2ww5" event={"ID":"2711a2e9-8d7a-4942-b762-f211d519e0f4","Type":"ContainerStarted","Data":"22ab8b66d0c971861f950fe20c987d304c035b5e427043bff1d53d23a6bf5a71"} Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.588514 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=32.588496428 podStartE2EDuration="32.588496428s" podCreationTimestamp="2025-10-03 13:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:12:46.585717861 +0000 UTC m=+1302.384190045" watchObservedRunningTime="2025-10-03 13:12:46.588496428 +0000 UTC m=+1302.386968612" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.608560 4578 scope.go:117] "RemoveContainer" containerID="b87bf43c564aa0714f110549b69b1d0bf12d6c2035f2dcecb4144e9b98bf2b9b" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.667284 4578 scope.go:117] "RemoveContainer" containerID="313818f2faab0b600157582f66168ad9ab8c897bda250d026f7b4356128816d1" Oct 03 13:12:46 crc kubenswrapper[4578]: E1003 13:12:46.667850 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"313818f2faab0b600157582f66168ad9ab8c897bda250d026f7b4356128816d1\": container with ID starting with 313818f2faab0b600157582f66168ad9ab8c897bda250d026f7b4356128816d1 not found: ID does not exist" containerID="313818f2faab0b600157582f66168ad9ab8c897bda250d026f7b4356128816d1" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.667907 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"313818f2faab0b600157582f66168ad9ab8c897bda250d026f7b4356128816d1"} err="failed to get container status \"313818f2faab0b600157582f66168ad9ab8c897bda250d026f7b4356128816d1\": rpc error: code = NotFound desc = could not find container \"313818f2faab0b600157582f66168ad9ab8c897bda250d026f7b4356128816d1\": container with ID starting with 313818f2faab0b600157582f66168ad9ab8c897bda250d026f7b4356128816d1 not found: ID does not exist" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.667935 4578 scope.go:117] "RemoveContainer" containerID="b87bf43c564aa0714f110549b69b1d0bf12d6c2035f2dcecb4144e9b98bf2b9b" Oct 03 13:12:46 crc kubenswrapper[4578]: E1003 13:12:46.669475 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b87bf43c564aa0714f110549b69b1d0bf12d6c2035f2dcecb4144e9b98bf2b9b\": container with ID starting with b87bf43c564aa0714f110549b69b1d0bf12d6c2035f2dcecb4144e9b98bf2b9b not found: ID does not exist" containerID="b87bf43c564aa0714f110549b69b1d0bf12d6c2035f2dcecb4144e9b98bf2b9b" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.669507 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b87bf43c564aa0714f110549b69b1d0bf12d6c2035f2dcecb4144e9b98bf2b9b"} err="failed to get container status \"b87bf43c564aa0714f110549b69b1d0bf12d6c2035f2dcecb4144e9b98bf2b9b\": rpc error: code = NotFound desc = could not find container \"b87bf43c564aa0714f110549b69b1d0bf12d6c2035f2dcecb4144e9b98bf2b9b\": container with ID starting with b87bf43c564aa0714f110549b69b1d0bf12d6c2035f2dcecb4144e9b98bf2b9b not found: ID does not exist" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.669529 4578 scope.go:117] "RemoveContainer" containerID="313818f2faab0b600157582f66168ad9ab8c897bda250d026f7b4356128816d1" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.670252 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"313818f2faab0b600157582f66168ad9ab8c897bda250d026f7b4356128816d1"} err="failed to get container status \"313818f2faab0b600157582f66168ad9ab8c897bda250d026f7b4356128816d1\": rpc error: code = NotFound desc = could not find container \"313818f2faab0b600157582f66168ad9ab8c897bda250d026f7b4356128816d1\": container with ID starting with 313818f2faab0b600157582f66168ad9ab8c897bda250d026f7b4356128816d1 not found: ID does not exist" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.670273 4578 scope.go:117] "RemoveContainer" containerID="b87bf43c564aa0714f110549b69b1d0bf12d6c2035f2dcecb4144e9b98bf2b9b" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.671774 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b87bf43c564aa0714f110549b69b1d0bf12d6c2035f2dcecb4144e9b98bf2b9b"} err="failed to get container status \"b87bf43c564aa0714f110549b69b1d0bf12d6c2035f2dcecb4144e9b98bf2b9b\": rpc error: code = NotFound desc = could not find container \"b87bf43c564aa0714f110549b69b1d0bf12d6c2035f2dcecb4144e9b98bf2b9b\": container with ID starting with b87bf43c564aa0714f110549b69b1d0bf12d6c2035f2dcecb4144e9b98bf2b9b not found: ID does not exist" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.780181 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.781118 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.889022 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.896966 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.929605 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80" path="/var/lib/kubelet/pods/3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80/volumes" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.930228 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:12:46 crc kubenswrapper[4578]: E1003 13:12:46.930494 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80" containerName="glance-httpd" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.930510 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80" containerName="glance-httpd" Oct 03 13:12:46 crc kubenswrapper[4578]: E1003 13:12:46.930546 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80" containerName="glance-log" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.930552 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80" containerName="glance-log" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.930743 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80" containerName="glance-log" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.930771 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="3beaf9af-dc3d-4cb6-91db-c3a68b7f6c80" containerName="glance-httpd" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.931962 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.934844 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.939080 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.939282 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.953602 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:46 crc kubenswrapper[4578]: I1003 13:12:46.953656 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.063069 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-logs\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.063498 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.063565 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-scripts\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.063581 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s8bm\" (UniqueName: \"kubernetes.io/projected/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-kube-api-access-7s8bm\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.063610 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.063696 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.063727 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.063777 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-config-data\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.164942 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-logs\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.165037 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.165089 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-scripts\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.165108 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s8bm\" (UniqueName: \"kubernetes.io/projected/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-kube-api-access-7s8bm\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.165135 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.165174 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.165196 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.165239 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-config-data\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.166489 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-logs\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.171893 4578 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.171990 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.172992 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.174261 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.187245 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-scripts\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.196250 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s8bm\" (UniqueName: \"kubernetes.io/projected/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-kube-api-access-7s8bm\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.208212 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-config-data\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.226973 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.269161 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.429996 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.585553 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-config-data\") pod \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.585586 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-scripts\") pod \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.585669 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-logs\") pod \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.585732 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-internal-tls-certs\") pod \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.585776 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fp6v\" (UniqueName: \"kubernetes.io/projected/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-kube-api-access-4fp6v\") pod \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.585828 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.585847 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-combined-ca-bundle\") pod \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.585863 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-httpd-run\") pod \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\" (UID: \"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5\") " Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.586978 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "fd6a1a68-a60c-41d3-a0e9-ca51feef6df5" (UID: "fd6a1a68-a60c-41d3-a0e9-ca51feef6df5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.587194 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-logs" (OuterVolumeSpecName: "logs") pod "fd6a1a68-a60c-41d3-a0e9-ca51feef6df5" (UID: "fd6a1a68-a60c-41d3-a0e9-ca51feef6df5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.599215 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "fd6a1a68-a60c-41d3-a0e9-ca51feef6df5" (UID: "fd6a1a68-a60c-41d3-a0e9-ca51feef6df5"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.599223 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-scripts" (OuterVolumeSpecName: "scripts") pod "fd6a1a68-a60c-41d3-a0e9-ca51feef6df5" (UID: "fd6a1a68-a60c-41d3-a0e9-ca51feef6df5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.605830 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-kube-api-access-4fp6v" (OuterVolumeSpecName: "kube-api-access-4fp6v") pod "fd6a1a68-a60c-41d3-a0e9-ca51feef6df5" (UID: "fd6a1a68-a60c-41d3-a0e9-ca51feef6df5"). InnerVolumeSpecName "kube-api-access-4fp6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.631145 4578 generic.go:334] "Generic (PLEG): container finished" podID="fd6a1a68-a60c-41d3-a0e9-ca51feef6df5" containerID="4204fc6b484771af3e2d7cbbbd547b7f401dba13ac6e8bcf7747e91bc1d2b151" exitCode=143 Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.631181 4578 generic.go:334] "Generic (PLEG): container finished" podID="fd6a1a68-a60c-41d3-a0e9-ca51feef6df5" containerID="c4550bc1dabcb2df378a1554b302e86b33ee5bc061cd46dc7496c1d0887a7210" exitCode=143 Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.631264 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5","Type":"ContainerDied","Data":"4204fc6b484771af3e2d7cbbbd547b7f401dba13ac6e8bcf7747e91bc1d2b151"} Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.631294 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5","Type":"ContainerDied","Data":"c4550bc1dabcb2df378a1554b302e86b33ee5bc061cd46dc7496c1d0887a7210"} Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.631311 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fd6a1a68-a60c-41d3-a0e9-ca51feef6df5","Type":"ContainerDied","Data":"be1412f89b62042a047caf0003a928e2c4fc004839bcc7cf4efc9425bc025dc8"} Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.631329 4578 scope.go:117] "RemoveContainer" containerID="4204fc6b484771af3e2d7cbbbd547b7f401dba13ac6e8bcf7747e91bc1d2b151" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.631464 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.641079 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd6a1a68-a60c-41d3-a0e9-ca51feef6df5" (UID: "fd6a1a68-a60c-41d3-a0e9-ca51feef6df5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.655077 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d7d5c9485-l2ww5" event={"ID":"2711a2e9-8d7a-4942-b762-f211d519e0f4","Type":"ContainerStarted","Data":"32b0703b15acfba9df59340edadd4f32b2cd336bddffaf9cfa8782ccf5bf0fdb"} Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.655139 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d7d5c9485-l2ww5" event={"ID":"2711a2e9-8d7a-4942-b762-f211d519e0f4","Type":"ContainerStarted","Data":"564799ba218e4dbb30a6bb231535273672ef82ed3cbcb8089c5a81ee6e7cd679"} Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.656662 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.691012 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fp6v\" (UniqueName: \"kubernetes.io/projected/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-kube-api-access-4fp6v\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.691065 4578 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.691080 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.691091 4578 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.691102 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.691115 4578 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.712875 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-d7d5c9485-l2ww5" podStartSLOduration=3.712844365 podStartE2EDuration="3.712844365s" podCreationTimestamp="2025-10-03 13:12:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:12:47.694545255 +0000 UTC m=+1303.493017439" watchObservedRunningTime="2025-10-03 13:12:47.712844365 +0000 UTC m=+1303.511316549" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.744291 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-config-data" (OuterVolumeSpecName: "config-data") pod "fd6a1a68-a60c-41d3-a0e9-ca51feef6df5" (UID: "fd6a1a68-a60c-41d3-a0e9-ca51feef6df5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.756186 4578 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.775886 4578 scope.go:117] "RemoveContainer" containerID="c4550bc1dabcb2df378a1554b302e86b33ee5bc061cd46dc7496c1d0887a7210" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.792366 4578 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.792400 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.818567 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "fd6a1a68-a60c-41d3-a0e9-ca51feef6df5" (UID: "fd6a1a68-a60c-41d3-a0e9-ca51feef6df5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.894247 4578 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.956588 4578 scope.go:117] "RemoveContainer" containerID="4204fc6b484771af3e2d7cbbbd547b7f401dba13ac6e8bcf7747e91bc1d2b151" Oct 03 13:12:47 crc kubenswrapper[4578]: E1003 13:12:47.957245 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4204fc6b484771af3e2d7cbbbd547b7f401dba13ac6e8bcf7747e91bc1d2b151\": container with ID starting with 4204fc6b484771af3e2d7cbbbd547b7f401dba13ac6e8bcf7747e91bc1d2b151 not found: ID does not exist" containerID="4204fc6b484771af3e2d7cbbbd547b7f401dba13ac6e8bcf7747e91bc1d2b151" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.957276 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4204fc6b484771af3e2d7cbbbd547b7f401dba13ac6e8bcf7747e91bc1d2b151"} err="failed to get container status \"4204fc6b484771af3e2d7cbbbd547b7f401dba13ac6e8bcf7747e91bc1d2b151\": rpc error: code = NotFound desc = could not find container \"4204fc6b484771af3e2d7cbbbd547b7f401dba13ac6e8bcf7747e91bc1d2b151\": container with ID starting with 4204fc6b484771af3e2d7cbbbd547b7f401dba13ac6e8bcf7747e91bc1d2b151 not found: ID does not exist" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.957308 4578 scope.go:117] "RemoveContainer" containerID="c4550bc1dabcb2df378a1554b302e86b33ee5bc061cd46dc7496c1d0887a7210" Oct 03 13:12:47 crc kubenswrapper[4578]: E1003 13:12:47.957741 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4550bc1dabcb2df378a1554b302e86b33ee5bc061cd46dc7496c1d0887a7210\": container with ID starting with c4550bc1dabcb2df378a1554b302e86b33ee5bc061cd46dc7496c1d0887a7210 not found: ID does not exist" containerID="c4550bc1dabcb2df378a1554b302e86b33ee5bc061cd46dc7496c1d0887a7210" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.957777 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4550bc1dabcb2df378a1554b302e86b33ee5bc061cd46dc7496c1d0887a7210"} err="failed to get container status \"c4550bc1dabcb2df378a1554b302e86b33ee5bc061cd46dc7496c1d0887a7210\": rpc error: code = NotFound desc = could not find container \"c4550bc1dabcb2df378a1554b302e86b33ee5bc061cd46dc7496c1d0887a7210\": container with ID starting with c4550bc1dabcb2df378a1554b302e86b33ee5bc061cd46dc7496c1d0887a7210 not found: ID does not exist" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.957801 4578 scope.go:117] "RemoveContainer" containerID="4204fc6b484771af3e2d7cbbbd547b7f401dba13ac6e8bcf7747e91bc1d2b151" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.961771 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4204fc6b484771af3e2d7cbbbd547b7f401dba13ac6e8bcf7747e91bc1d2b151"} err="failed to get container status \"4204fc6b484771af3e2d7cbbbd547b7f401dba13ac6e8bcf7747e91bc1d2b151\": rpc error: code = NotFound desc = could not find container \"4204fc6b484771af3e2d7cbbbd547b7f401dba13ac6e8bcf7747e91bc1d2b151\": container with ID starting with 4204fc6b484771af3e2d7cbbbd547b7f401dba13ac6e8bcf7747e91bc1d2b151 not found: ID does not exist" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.961794 4578 scope.go:117] "RemoveContainer" containerID="c4550bc1dabcb2df378a1554b302e86b33ee5bc061cd46dc7496c1d0887a7210" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.966652 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4550bc1dabcb2df378a1554b302e86b33ee5bc061cd46dc7496c1d0887a7210"} err="failed to get container status \"c4550bc1dabcb2df378a1554b302e86b33ee5bc061cd46dc7496c1d0887a7210\": rpc error: code = NotFound desc = could not find container \"c4550bc1dabcb2df378a1554b302e86b33ee5bc061cd46dc7496c1d0887a7210\": container with ID starting with c4550bc1dabcb2df378a1554b302e86b33ee5bc061cd46dc7496c1d0887a7210 not found: ID does not exist" Oct 03 13:12:47 crc kubenswrapper[4578]: I1003 13:12:47.979844 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.005365 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.022721 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:12:48 crc kubenswrapper[4578]: E1003 13:12:48.023140 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd6a1a68-a60c-41d3-a0e9-ca51feef6df5" containerName="glance-httpd" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.023153 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd6a1a68-a60c-41d3-a0e9-ca51feef6df5" containerName="glance-httpd" Oct 03 13:12:48 crc kubenswrapper[4578]: E1003 13:12:48.023174 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd6a1a68-a60c-41d3-a0e9-ca51feef6df5" containerName="glance-log" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.023181 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd6a1a68-a60c-41d3-a0e9-ca51feef6df5" containerName="glance-log" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.023354 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd6a1a68-a60c-41d3-a0e9-ca51feef6df5" containerName="glance-httpd" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.023373 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd6a1a68-a60c-41d3-a0e9-ca51feef6df5" containerName="glance-log" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.024267 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.027662 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.033276 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.121380 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.216117 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-config-data\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.216189 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25ef994a-0092-4e72-a636-f51bbd1a8a99-logs\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.216236 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.216313 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.216336 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/25ef994a-0092-4e72-a636-f51bbd1a8a99-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.216399 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-scripts\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.216432 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gjc7\" (UniqueName: \"kubernetes.io/projected/25ef994a-0092-4e72-a636-f51bbd1a8a99-kube-api-access-5gjc7\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.216530 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.221847 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.318742 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25ef994a-0092-4e72-a636-f51bbd1a8a99-logs\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.319172 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.319272 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25ef994a-0092-4e72-a636-f51bbd1a8a99-logs\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.319377 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.319466 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/25ef994a-0092-4e72-a636-f51bbd1a8a99-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.319596 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-scripts\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.319716 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gjc7\" (UniqueName: \"kubernetes.io/projected/25ef994a-0092-4e72-a636-f51bbd1a8a99-kube-api-access-5gjc7\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.319834 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.319937 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-config-data\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.319605 4578 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.320758 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/25ef994a-0092-4e72-a636-f51bbd1a8a99-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.329562 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-config-data\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.346667 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-scripts\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.350979 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.352434 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.354534 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gjc7\" (UniqueName: \"kubernetes.io/projected/25ef994a-0092-4e72-a636-f51bbd1a8a99-kube-api-access-5gjc7\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.368223 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.652106 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.693804 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3","Type":"ContainerStarted","Data":"c84cdde022f3813331f17ad5d123c3f66050af65f2e5bc51d7a22a5c7a232e82"} Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.698807 4578 generic.go:334] "Generic (PLEG): container finished" podID="eb6557a6-8665-4a81-afec-4ccaef43b801" containerID="d3deb53df311f3445d8b678731a0915eeb05d092fbb84ef502995999cad84baf" exitCode=0 Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.699615 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gv9w6" event={"ID":"eb6557a6-8665-4a81-afec-4ccaef43b801","Type":"ContainerDied","Data":"d3deb53df311f3445d8b678731a0915eeb05d092fbb84ef502995999cad84baf"} Oct 03 13:12:48 crc kubenswrapper[4578]: I1003 13:12:48.931921 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd6a1a68-a60c-41d3-a0e9-ca51feef6df5" path="/var/lib/kubelet/pods/fd6a1a68-a60c-41d3-a0e9-ca51feef6df5/volumes" Oct 03 13:12:49 crc kubenswrapper[4578]: I1003 13:12:49.398175 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:12:49 crc kubenswrapper[4578]: W1003 13:12:49.416114 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25ef994a_0092_4e72_a636_f51bbd1a8a99.slice/crio-cc772551e8d3098cae7f8cc16d31020fbd3c6af7e2c008694afe493d7a5e0cc1 WatchSource:0}: Error finding container cc772551e8d3098cae7f8cc16d31020fbd3c6af7e2c008694afe493d7a5e0cc1: Status 404 returned error can't find the container with id cc772551e8d3098cae7f8cc16d31020fbd3c6af7e2c008694afe493d7a5e0cc1 Oct 03 13:12:49 crc kubenswrapper[4578]: I1003 13:12:49.757705 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"25ef994a-0092-4e72-a636-f51bbd1a8a99","Type":"ContainerStarted","Data":"cc772551e8d3098cae7f8cc16d31020fbd3c6af7e2c008694afe493d7a5e0cc1"} Oct 03 13:12:49 crc kubenswrapper[4578]: I1003 13:12:49.759525 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3","Type":"ContainerStarted","Data":"44e11f2ae23fddbb737f7b00fdf3e82e831289bd9ebbaf9260098bbe8900732b"} Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.382451 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gv9w6" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.567290 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb6557a6-8665-4a81-afec-4ccaef43b801-combined-ca-bundle\") pod \"eb6557a6-8665-4a81-afec-4ccaef43b801\" (UID: \"eb6557a6-8665-4a81-afec-4ccaef43b801\") " Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.567679 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb6557a6-8665-4a81-afec-4ccaef43b801-logs\") pod \"eb6557a6-8665-4a81-afec-4ccaef43b801\" (UID: \"eb6557a6-8665-4a81-afec-4ccaef43b801\") " Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.567795 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb6557a6-8665-4a81-afec-4ccaef43b801-config-data\") pod \"eb6557a6-8665-4a81-afec-4ccaef43b801\" (UID: \"eb6557a6-8665-4a81-afec-4ccaef43b801\") " Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.567933 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb6557a6-8665-4a81-afec-4ccaef43b801-scripts\") pod \"eb6557a6-8665-4a81-afec-4ccaef43b801\" (UID: \"eb6557a6-8665-4a81-afec-4ccaef43b801\") " Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.567967 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hph4p\" (UniqueName: \"kubernetes.io/projected/eb6557a6-8665-4a81-afec-4ccaef43b801-kube-api-access-hph4p\") pod \"eb6557a6-8665-4a81-afec-4ccaef43b801\" (UID: \"eb6557a6-8665-4a81-afec-4ccaef43b801\") " Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.583170 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb6557a6-8665-4a81-afec-4ccaef43b801-logs" (OuterVolumeSpecName: "logs") pod "eb6557a6-8665-4a81-afec-4ccaef43b801" (UID: "eb6557a6-8665-4a81-afec-4ccaef43b801"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.583587 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb6557a6-8665-4a81-afec-4ccaef43b801-kube-api-access-hph4p" (OuterVolumeSpecName: "kube-api-access-hph4p") pod "eb6557a6-8665-4a81-afec-4ccaef43b801" (UID: "eb6557a6-8665-4a81-afec-4ccaef43b801"). InnerVolumeSpecName "kube-api-access-hph4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.591731 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb6557a6-8665-4a81-afec-4ccaef43b801-scripts" (OuterVolumeSpecName: "scripts") pod "eb6557a6-8665-4a81-afec-4ccaef43b801" (UID: "eb6557a6-8665-4a81-afec-4ccaef43b801"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.608962 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb6557a6-8665-4a81-afec-4ccaef43b801-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb6557a6-8665-4a81-afec-4ccaef43b801" (UID: "eb6557a6-8665-4a81-afec-4ccaef43b801"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.629674 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-d5f486769-rxsds" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.658238 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb6557a6-8665-4a81-afec-4ccaef43b801-config-data" (OuterVolumeSpecName: "config-data") pod "eb6557a6-8665-4a81-afec-4ccaef43b801" (UID: "eb6557a6-8665-4a81-afec-4ccaef43b801"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.670372 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb6557a6-8665-4a81-afec-4ccaef43b801-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.670406 4578 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb6557a6-8665-4a81-afec-4ccaef43b801-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.670416 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb6557a6-8665-4a81-afec-4ccaef43b801-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.670424 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb6557a6-8665-4a81-afec-4ccaef43b801-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.670435 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hph4p\" (UniqueName: \"kubernetes.io/projected/eb6557a6-8665-4a81-afec-4ccaef43b801-kube-api-access-hph4p\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.777419 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"25ef994a-0092-4e72-a636-f51bbd1a8a99","Type":"ContainerStarted","Data":"666b5fd3073ef524529191139ec8cb50cfbb3d1d8b1a4b19cea1ea993e443544"} Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.797812 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3","Type":"ContainerStarted","Data":"930e4b153894449384f40e1fcaa46f1750f077197fa07f10a8b4af27ab85795a"} Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.805391 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gv9w6" event={"ID":"eb6557a6-8665-4a81-afec-4ccaef43b801","Type":"ContainerDied","Data":"65ea7c7a946a2329012a28d8b4a6426be4cbb51ee7f3f5028c004af036781f12"} Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.805443 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65ea7c7a946a2329012a28d8b4a6426be4cbb51ee7f3f5028c004af036781f12" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.805521 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gv9w6" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.871397 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.871372292 podStartE2EDuration="4.871372292s" podCreationTimestamp="2025-10-03 13:12:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:12:50.832323256 +0000 UTC m=+1306.630795440" watchObservedRunningTime="2025-10-03 13:12:50.871372292 +0000 UTC m=+1306.669844466" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.848471 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-66cbcb46d8-9jksg"] Oct 03 13:12:50 crc kubenswrapper[4578]: E1003 13:12:50.883837 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb6557a6-8665-4a81-afec-4ccaef43b801" containerName="placement-db-sync" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.883865 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb6557a6-8665-4a81-afec-4ccaef43b801" containerName="placement-db-sync" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.884149 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb6557a6-8665-4a81-afec-4ccaef43b801" containerName="placement-db-sync" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.885220 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.890255 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-66cbcb46d8-9jksg"] Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.891426 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.893478 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.893519 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.893560 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-d26lq" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.893612 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.975461 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-logs\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.975820 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj774\" (UniqueName: \"kubernetes.io/projected/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-kube-api-access-bj774\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.975856 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-scripts\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.975896 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-config-data\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.975930 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-internal-tls-certs\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.975978 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-combined-ca-bundle\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:50 crc kubenswrapper[4578]: I1003 13:12:50.975994 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-public-tls-certs\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.078481 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-logs\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.078549 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj774\" (UniqueName: \"kubernetes.io/projected/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-kube-api-access-bj774\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.078589 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-scripts\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.078615 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-config-data\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.078660 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-internal-tls-certs\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.078698 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-combined-ca-bundle\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.078920 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-logs\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.079023 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-public-tls-certs\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.086076 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-scripts\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.086222 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-combined-ca-bundle\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.087095 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-internal-tls-certs\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.089026 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-public-tls-certs\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.098037 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-config-data\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.099330 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj774\" (UniqueName: \"kubernetes.io/projected/5e89457e-7c3d-4f85-a14d-ca47f5bdca07-kube-api-access-bj774\") pod \"placement-66cbcb46d8-9jksg\" (UID: \"5e89457e-7c3d-4f85-a14d-ca47f5bdca07\") " pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.206849 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.238859 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.298490 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-2gwvm"] Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.298779 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" podUID="3579029a-5cd2-43d7-9611-7e77a76acd5c" containerName="dnsmasq-dns" containerID="cri-o://5e3db199e76de0240c635a99e2ecc1c9e01f4c85de1237bb45f06ad520a825eb" gracePeriod=10 Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.823818 4578 generic.go:334] "Generic (PLEG): container finished" podID="eacf42c3-8e6f-4304-b17b-de32359e8d76" containerID="5fa1c518576e9d2ffc9dda2b0ef5adaf2cf39618bda67a1eee0fb45f6bd0f0f4" exitCode=0 Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.824135 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-j9scz" event={"ID":"eacf42c3-8e6f-4304-b17b-de32359e8d76","Type":"ContainerDied","Data":"5fa1c518576e9d2ffc9dda2b0ef5adaf2cf39618bda67a1eee0fb45f6bd0f0f4"} Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.842829 4578 generic.go:334] "Generic (PLEG): container finished" podID="3579029a-5cd2-43d7-9611-7e77a76acd5c" containerID="5e3db199e76de0240c635a99e2ecc1c9e01f4c85de1237bb45f06ad520a825eb" exitCode=0 Oct 03 13:12:51 crc kubenswrapper[4578]: I1003 13:12:51.842927 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" event={"ID":"3579029a-5cd2-43d7-9611-7e77a76acd5c","Type":"ContainerDied","Data":"5e3db199e76de0240c635a99e2ecc1c9e01f4c85de1237bb45f06ad520a825eb"} Oct 03 13:12:52 crc kubenswrapper[4578]: I1003 13:12:52.759851 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" podUID="3579029a-5cd2-43d7-9611-7e77a76acd5c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: connect: connection refused" Oct 03 13:12:53 crc kubenswrapper[4578]: I1003 13:12:53.865060 4578 generic.go:334] "Generic (PLEG): container finished" podID="a9856a93-43ca-4f3d-96a5-50c85086ae91" containerID="4c31f2c7bec5ec73b4641d8ab5761d2f40e161d8db396a1f5ffe89700025a8c9" exitCode=0 Oct 03 13:12:53 crc kubenswrapper[4578]: I1003 13:12:53.865099 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jrdpv" event={"ID":"a9856a93-43ca-4f3d-96a5-50c85086ae91","Type":"ContainerDied","Data":"4c31f2c7bec5ec73b4641d8ab5761d2f40e161d8db396a1f5ffe89700025a8c9"} Oct 03 13:12:54 crc kubenswrapper[4578]: I1003 13:12:54.456339 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-j9scz" Oct 03 13:12:54 crc kubenswrapper[4578]: I1003 13:12:54.557323 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eacf42c3-8e6f-4304-b17b-de32359e8d76-combined-ca-bundle\") pod \"eacf42c3-8e6f-4304-b17b-de32359e8d76\" (UID: \"eacf42c3-8e6f-4304-b17b-de32359e8d76\") " Oct 03 13:12:54 crc kubenswrapper[4578]: I1003 13:12:54.557833 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9lsz\" (UniqueName: \"kubernetes.io/projected/eacf42c3-8e6f-4304-b17b-de32359e8d76-kube-api-access-j9lsz\") pod \"eacf42c3-8e6f-4304-b17b-de32359e8d76\" (UID: \"eacf42c3-8e6f-4304-b17b-de32359e8d76\") " Oct 03 13:12:54 crc kubenswrapper[4578]: I1003 13:12:54.557925 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eacf42c3-8e6f-4304-b17b-de32359e8d76-db-sync-config-data\") pod \"eacf42c3-8e6f-4304-b17b-de32359e8d76\" (UID: \"eacf42c3-8e6f-4304-b17b-de32359e8d76\") " Oct 03 13:12:54 crc kubenswrapper[4578]: I1003 13:12:54.566865 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eacf42c3-8e6f-4304-b17b-de32359e8d76-kube-api-access-j9lsz" (OuterVolumeSpecName: "kube-api-access-j9lsz") pod "eacf42c3-8e6f-4304-b17b-de32359e8d76" (UID: "eacf42c3-8e6f-4304-b17b-de32359e8d76"). InnerVolumeSpecName "kube-api-access-j9lsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:12:54 crc kubenswrapper[4578]: I1003 13:12:54.611780 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eacf42c3-8e6f-4304-b17b-de32359e8d76-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "eacf42c3-8e6f-4304-b17b-de32359e8d76" (UID: "eacf42c3-8e6f-4304-b17b-de32359e8d76"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:54 crc kubenswrapper[4578]: I1003 13:12:54.654900 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eacf42c3-8e6f-4304-b17b-de32359e8d76-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eacf42c3-8e6f-4304-b17b-de32359e8d76" (UID: "eacf42c3-8e6f-4304-b17b-de32359e8d76"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:54 crc kubenswrapper[4578]: I1003 13:12:54.659986 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9lsz\" (UniqueName: \"kubernetes.io/projected/eacf42c3-8e6f-4304-b17b-de32359e8d76-kube-api-access-j9lsz\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:54 crc kubenswrapper[4578]: I1003 13:12:54.660120 4578 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eacf42c3-8e6f-4304-b17b-de32359e8d76-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:54 crc kubenswrapper[4578]: I1003 13:12:54.660179 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eacf42c3-8e6f-4304-b17b-de32359e8d76-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:54 crc kubenswrapper[4578]: I1003 13:12:54.876129 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-j9scz" Oct 03 13:12:54 crc kubenswrapper[4578]: I1003 13:12:54.876137 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-j9scz" event={"ID":"eacf42c3-8e6f-4304-b17b-de32359e8d76","Type":"ContainerDied","Data":"e8a11b5a71783ef74f5b46f522b223afb8469a4e382eeb0af5f7b109d4090034"} Oct 03 13:12:54 crc kubenswrapper[4578]: I1003 13:12:54.878535 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8a11b5a71783ef74f5b46f522b223afb8469a4e382eeb0af5f7b109d4090034" Oct 03 13:12:55 crc kubenswrapper[4578]: I1003 13:12:55.878783 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-d9f65c5d7-qf4l9"] Oct 03 13:12:55 crc kubenswrapper[4578]: E1003 13:12:55.879121 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eacf42c3-8e6f-4304-b17b-de32359e8d76" containerName="barbican-db-sync" Oct 03 13:12:55 crc kubenswrapper[4578]: I1003 13:12:55.879134 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="eacf42c3-8e6f-4304-b17b-de32359e8d76" containerName="barbican-db-sync" Oct 03 13:12:55 crc kubenswrapper[4578]: I1003 13:12:55.879328 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="eacf42c3-8e6f-4304-b17b-de32359e8d76" containerName="barbican-db-sync" Oct 03 13:12:55 crc kubenswrapper[4578]: I1003 13:12:55.885449 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-d9f65c5d7-qf4l9" Oct 03 13:12:55 crc kubenswrapper[4578]: I1003 13:12:55.918938 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8fxh8" Oct 03 13:12:55 crc kubenswrapper[4578]: I1003 13:12:55.919389 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 03 13:12:55 crc kubenswrapper[4578]: I1003 13:12:55.935358 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-d9f65c5d7-qf4l9"] Oct 03 13:12:55 crc kubenswrapper[4578]: I1003 13:12:55.960355 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 03 13:12:55 crc kubenswrapper[4578]: I1003 13:12:55.967955 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-786f64db8b-9n26w"] Oct 03 13:12:55 crc kubenswrapper[4578]: I1003 13:12:55.974706 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" Oct 03 13:12:55 crc kubenswrapper[4578]: I1003 13:12:55.986562 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 03 13:12:55 crc kubenswrapper[4578]: I1003 13:12:55.987447 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c187292d-a86d-455d-9c52-dabc5c3b4631-config-data\") pod \"barbican-worker-d9f65c5d7-qf4l9\" (UID: \"c187292d-a86d-455d-9c52-dabc5c3b4631\") " pod="openstack/barbican-worker-d9f65c5d7-qf4l9" Oct 03 13:12:55 crc kubenswrapper[4578]: I1003 13:12:55.987501 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c187292d-a86d-455d-9c52-dabc5c3b4631-logs\") pod \"barbican-worker-d9f65c5d7-qf4l9\" (UID: \"c187292d-a86d-455d-9c52-dabc5c3b4631\") " pod="openstack/barbican-worker-d9f65c5d7-qf4l9" Oct 03 13:12:55 crc kubenswrapper[4578]: I1003 13:12:55.987541 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js9lq\" (UniqueName: \"kubernetes.io/projected/c187292d-a86d-455d-9c52-dabc5c3b4631-kube-api-access-js9lq\") pod \"barbican-worker-d9f65c5d7-qf4l9\" (UID: \"c187292d-a86d-455d-9c52-dabc5c3b4631\") " pod="openstack/barbican-worker-d9f65c5d7-qf4l9" Oct 03 13:12:55 crc kubenswrapper[4578]: I1003 13:12:55.987644 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c187292d-a86d-455d-9c52-dabc5c3b4631-config-data-custom\") pod \"barbican-worker-d9f65c5d7-qf4l9\" (UID: \"c187292d-a86d-455d-9c52-dabc5c3b4631\") " pod="openstack/barbican-worker-d9f65c5d7-qf4l9" Oct 03 13:12:55 crc kubenswrapper[4578]: I1003 13:12:55.987709 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c187292d-a86d-455d-9c52-dabc5c3b4631-combined-ca-bundle\") pod \"barbican-worker-d9f65c5d7-qf4l9\" (UID: \"c187292d-a86d-455d-9c52-dabc5c3b4631\") " pod="openstack/barbican-worker-d9f65c5d7-qf4l9" Oct 03 13:12:55 crc kubenswrapper[4578]: I1003 13:12:55.997586 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-786f64db8b-9n26w"] Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.089694 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p6pq\" (UniqueName: \"kubernetes.io/projected/3bb09a8d-f51b-494d-8a75-8690e842a916-kube-api-access-4p6pq\") pod \"barbican-keystone-listener-786f64db8b-9n26w\" (UID: \"3bb09a8d-f51b-494d-8a75-8690e842a916\") " pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.089749 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bb09a8d-f51b-494d-8a75-8690e842a916-config-data\") pod \"barbican-keystone-listener-786f64db8b-9n26w\" (UID: \"3bb09a8d-f51b-494d-8a75-8690e842a916\") " pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.089776 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c187292d-a86d-455d-9c52-dabc5c3b4631-config-data\") pod \"barbican-worker-d9f65c5d7-qf4l9\" (UID: \"c187292d-a86d-455d-9c52-dabc5c3b4631\") " pod="openstack/barbican-worker-d9f65c5d7-qf4l9" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.089806 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c187292d-a86d-455d-9c52-dabc5c3b4631-logs\") pod \"barbican-worker-d9f65c5d7-qf4l9\" (UID: \"c187292d-a86d-455d-9c52-dabc5c3b4631\") " pod="openstack/barbican-worker-d9f65c5d7-qf4l9" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.089839 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js9lq\" (UniqueName: \"kubernetes.io/projected/c187292d-a86d-455d-9c52-dabc5c3b4631-kube-api-access-js9lq\") pod \"barbican-worker-d9f65c5d7-qf4l9\" (UID: \"c187292d-a86d-455d-9c52-dabc5c3b4631\") " pod="openstack/barbican-worker-d9f65c5d7-qf4l9" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.089884 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c187292d-a86d-455d-9c52-dabc5c3b4631-config-data-custom\") pod \"barbican-worker-d9f65c5d7-qf4l9\" (UID: \"c187292d-a86d-455d-9c52-dabc5c3b4631\") " pod="openstack/barbican-worker-d9f65c5d7-qf4l9" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.089907 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bb09a8d-f51b-494d-8a75-8690e842a916-combined-ca-bundle\") pod \"barbican-keystone-listener-786f64db8b-9n26w\" (UID: \"3bb09a8d-f51b-494d-8a75-8690e842a916\") " pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.089923 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3bb09a8d-f51b-494d-8a75-8690e842a916-config-data-custom\") pod \"barbican-keystone-listener-786f64db8b-9n26w\" (UID: \"3bb09a8d-f51b-494d-8a75-8690e842a916\") " pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.089950 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bb09a8d-f51b-494d-8a75-8690e842a916-logs\") pod \"barbican-keystone-listener-786f64db8b-9n26w\" (UID: \"3bb09a8d-f51b-494d-8a75-8690e842a916\") " pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.089985 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c187292d-a86d-455d-9c52-dabc5c3b4631-combined-ca-bundle\") pod \"barbican-worker-d9f65c5d7-qf4l9\" (UID: \"c187292d-a86d-455d-9c52-dabc5c3b4631\") " pod="openstack/barbican-worker-d9f65c5d7-qf4l9" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.096594 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c187292d-a86d-455d-9c52-dabc5c3b4631-logs\") pod \"barbican-worker-d9f65c5d7-qf4l9\" (UID: \"c187292d-a86d-455d-9c52-dabc5c3b4631\") " pod="openstack/barbican-worker-d9f65c5d7-qf4l9" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.098678 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c187292d-a86d-455d-9c52-dabc5c3b4631-config-data\") pod \"barbican-worker-d9f65c5d7-qf4l9\" (UID: \"c187292d-a86d-455d-9c52-dabc5c3b4631\") " pod="openstack/barbican-worker-d9f65c5d7-qf4l9" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.106495 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c187292d-a86d-455d-9c52-dabc5c3b4631-combined-ca-bundle\") pod \"barbican-worker-d9f65c5d7-qf4l9\" (UID: \"c187292d-a86d-455d-9c52-dabc5c3b4631\") " pod="openstack/barbican-worker-d9f65c5d7-qf4l9" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.120360 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-x5hxs"] Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.121838 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.122728 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c187292d-a86d-455d-9c52-dabc5c3b4631-config-data-custom\") pod \"barbican-worker-d9f65c5d7-qf4l9\" (UID: \"c187292d-a86d-455d-9c52-dabc5c3b4631\") " pod="openstack/barbican-worker-d9f65c5d7-qf4l9" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.162380 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-x5hxs"] Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.187250 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js9lq\" (UniqueName: \"kubernetes.io/projected/c187292d-a86d-455d-9c52-dabc5c3b4631-kube-api-access-js9lq\") pod \"barbican-worker-d9f65c5d7-qf4l9\" (UID: \"c187292d-a86d-455d-9c52-dabc5c3b4631\") " pod="openstack/barbican-worker-d9f65c5d7-qf4l9" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.191737 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p6pq\" (UniqueName: \"kubernetes.io/projected/3bb09a8d-f51b-494d-8a75-8690e842a916-kube-api-access-4p6pq\") pod \"barbican-keystone-listener-786f64db8b-9n26w\" (UID: \"3bb09a8d-f51b-494d-8a75-8690e842a916\") " pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.191783 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-config\") pod \"dnsmasq-dns-75c8ddd69c-x5hxs\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.191814 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bb09a8d-f51b-494d-8a75-8690e842a916-config-data\") pod \"barbican-keystone-listener-786f64db8b-9n26w\" (UID: \"3bb09a8d-f51b-494d-8a75-8690e842a916\") " pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.191831 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-x5hxs\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.191860 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-x5hxs\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.191895 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-x5hxs\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.191925 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-x5hxs\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.191939 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdctb\" (UniqueName: \"kubernetes.io/projected/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-kube-api-access-sdctb\") pod \"dnsmasq-dns-75c8ddd69c-x5hxs\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.191963 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bb09a8d-f51b-494d-8a75-8690e842a916-combined-ca-bundle\") pod \"barbican-keystone-listener-786f64db8b-9n26w\" (UID: \"3bb09a8d-f51b-494d-8a75-8690e842a916\") " pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.191978 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3bb09a8d-f51b-494d-8a75-8690e842a916-config-data-custom\") pod \"barbican-keystone-listener-786f64db8b-9n26w\" (UID: \"3bb09a8d-f51b-494d-8a75-8690e842a916\") " pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.192002 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bb09a8d-f51b-494d-8a75-8690e842a916-logs\") pod \"barbican-keystone-listener-786f64db8b-9n26w\" (UID: \"3bb09a8d-f51b-494d-8a75-8690e842a916\") " pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.192398 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bb09a8d-f51b-494d-8a75-8690e842a916-logs\") pod \"barbican-keystone-listener-786f64db8b-9n26w\" (UID: \"3bb09a8d-f51b-494d-8a75-8690e842a916\") " pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.195804 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bb09a8d-f51b-494d-8a75-8690e842a916-combined-ca-bundle\") pod \"barbican-keystone-listener-786f64db8b-9n26w\" (UID: \"3bb09a8d-f51b-494d-8a75-8690e842a916\") " pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.196297 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bb09a8d-f51b-494d-8a75-8690e842a916-config-data\") pod \"barbican-keystone-listener-786f64db8b-9n26w\" (UID: \"3bb09a8d-f51b-494d-8a75-8690e842a916\") " pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.198536 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3bb09a8d-f51b-494d-8a75-8690e842a916-config-data-custom\") pod \"barbican-keystone-listener-786f64db8b-9n26w\" (UID: \"3bb09a8d-f51b-494d-8a75-8690e842a916\") " pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.236007 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-d9f65c5d7-qf4l9" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.261695 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p6pq\" (UniqueName: \"kubernetes.io/projected/3bb09a8d-f51b-494d-8a75-8690e842a916-kube-api-access-4p6pq\") pod \"barbican-keystone-listener-786f64db8b-9n26w\" (UID: \"3bb09a8d-f51b-494d-8a75-8690e842a916\") " pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.295971 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-config\") pod \"dnsmasq-dns-75c8ddd69c-x5hxs\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.296023 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-x5hxs\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.296059 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-x5hxs\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.296118 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-x5hxs\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.296150 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-x5hxs\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.296171 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdctb\" (UniqueName: \"kubernetes.io/projected/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-kube-api-access-sdctb\") pod \"dnsmasq-dns-75c8ddd69c-x5hxs\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.298290 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-config\") pod \"dnsmasq-dns-75c8ddd69c-x5hxs\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.309063 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-x5hxs\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.309595 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-x5hxs\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.309734 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-x5hxs\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.310330 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-x5hxs\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.315906 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.364884 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdctb\" (UniqueName: \"kubernetes.io/projected/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-kube-api-access-sdctb\") pod \"dnsmasq-dns-75c8ddd69c-x5hxs\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.522703 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-85d5675966-rb6nf"] Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.525261 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.531069 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-85d5675966-rb6nf"] Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.536172 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.604836 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b40b382f-7fb7-4f67-9237-e5148abf7610-config-data-custom\") pod \"barbican-api-85d5675966-rb6nf\" (UID: \"b40b382f-7fb7-4f67-9237-e5148abf7610\") " pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.604960 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b40b382f-7fb7-4f67-9237-e5148abf7610-combined-ca-bundle\") pod \"barbican-api-85d5675966-rb6nf\" (UID: \"b40b382f-7fb7-4f67-9237-e5148abf7610\") " pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.604997 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b40b382f-7fb7-4f67-9237-e5148abf7610-logs\") pod \"barbican-api-85d5675966-rb6nf\" (UID: \"b40b382f-7fb7-4f67-9237-e5148abf7610\") " pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.605021 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b40b382f-7fb7-4f67-9237-e5148abf7610-config-data\") pod \"barbican-api-85d5675966-rb6nf\" (UID: \"b40b382f-7fb7-4f67-9237-e5148abf7610\") " pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.605131 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffqmv\" (UniqueName: \"kubernetes.io/projected/b40b382f-7fb7-4f67-9237-e5148abf7610-kube-api-access-ffqmv\") pod \"barbican-api-85d5675966-rb6nf\" (UID: \"b40b382f-7fb7-4f67-9237-e5148abf7610\") " pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.644699 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.706639 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffqmv\" (UniqueName: \"kubernetes.io/projected/b40b382f-7fb7-4f67-9237-e5148abf7610-kube-api-access-ffqmv\") pod \"barbican-api-85d5675966-rb6nf\" (UID: \"b40b382f-7fb7-4f67-9237-e5148abf7610\") " pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.706735 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b40b382f-7fb7-4f67-9237-e5148abf7610-config-data-custom\") pod \"barbican-api-85d5675966-rb6nf\" (UID: \"b40b382f-7fb7-4f67-9237-e5148abf7610\") " pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.706761 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b40b382f-7fb7-4f67-9237-e5148abf7610-combined-ca-bundle\") pod \"barbican-api-85d5675966-rb6nf\" (UID: \"b40b382f-7fb7-4f67-9237-e5148abf7610\") " pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.706788 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b40b382f-7fb7-4f67-9237-e5148abf7610-logs\") pod \"barbican-api-85d5675966-rb6nf\" (UID: \"b40b382f-7fb7-4f67-9237-e5148abf7610\") " pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.706804 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b40b382f-7fb7-4f67-9237-e5148abf7610-config-data\") pod \"barbican-api-85d5675966-rb6nf\" (UID: \"b40b382f-7fb7-4f67-9237-e5148abf7610\") " pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.707948 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b40b382f-7fb7-4f67-9237-e5148abf7610-logs\") pod \"barbican-api-85d5675966-rb6nf\" (UID: \"b40b382f-7fb7-4f67-9237-e5148abf7610\") " pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.712746 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b40b382f-7fb7-4f67-9237-e5148abf7610-combined-ca-bundle\") pod \"barbican-api-85d5675966-rb6nf\" (UID: \"b40b382f-7fb7-4f67-9237-e5148abf7610\") " pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.717799 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b40b382f-7fb7-4f67-9237-e5148abf7610-config-data\") pod \"barbican-api-85d5675966-rb6nf\" (UID: \"b40b382f-7fb7-4f67-9237-e5148abf7610\") " pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.718552 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b40b382f-7fb7-4f67-9237-e5148abf7610-config-data-custom\") pod \"barbican-api-85d5675966-rb6nf\" (UID: \"b40b382f-7fb7-4f67-9237-e5148abf7610\") " pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.735241 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffqmv\" (UniqueName: \"kubernetes.io/projected/b40b382f-7fb7-4f67-9237-e5148abf7610-kube-api-access-ffqmv\") pod \"barbican-api-85d5675966-rb6nf\" (UID: \"b40b382f-7fb7-4f67-9237-e5148abf7610\") " pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.781611 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7b58cf9568-f8nng" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.856621 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:12:56 crc kubenswrapper[4578]: I1003 13:12:56.948459 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-68684899bd-2cwll" podUID="d49d623e-2cd6-4786-b1da-34b78b89a134" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 03 13:12:57 crc kubenswrapper[4578]: I1003 13:12:57.270403 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 03 13:12:57 crc kubenswrapper[4578]: I1003 13:12:57.270750 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 03 13:12:57 crc kubenswrapper[4578]: I1003 13:12:57.331103 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 03 13:12:57 crc kubenswrapper[4578]: I1003 13:12:57.346458 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 03 13:12:57 crc kubenswrapper[4578]: I1003 13:12:57.909022 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 03 13:12:57 crc kubenswrapper[4578]: I1003 13:12:57.909063 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.712890 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.771118 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-credential-keys\") pod \"a9856a93-43ca-4f3d-96a5-50c85086ae91\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.771168 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qq528\" (UniqueName: \"kubernetes.io/projected/a9856a93-43ca-4f3d-96a5-50c85086ae91-kube-api-access-qq528\") pod \"a9856a93-43ca-4f3d-96a5-50c85086ae91\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.771286 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-combined-ca-bundle\") pod \"a9856a93-43ca-4f3d-96a5-50c85086ae91\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.771315 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-config-data\") pod \"a9856a93-43ca-4f3d-96a5-50c85086ae91\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.771342 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-fernet-keys\") pod \"a9856a93-43ca-4f3d-96a5-50c85086ae91\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.771435 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-scripts\") pod \"a9856a93-43ca-4f3d-96a5-50c85086ae91\" (UID: \"a9856a93-43ca-4f3d-96a5-50c85086ae91\") " Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.773786 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-f87d658c4-wf95n"] Oct 03 13:12:59 crc kubenswrapper[4578]: E1003 13:12:59.774175 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9856a93-43ca-4f3d-96a5-50c85086ae91" containerName="keystone-bootstrap" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.774187 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9856a93-43ca-4f3d-96a5-50c85086ae91" containerName="keystone-bootstrap" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.774365 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9856a93-43ca-4f3d-96a5-50c85086ae91" containerName="keystone-bootstrap" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.775254 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.777573 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a9856a93-43ca-4f3d-96a5-50c85086ae91" (UID: "a9856a93-43ca-4f3d-96a5-50c85086ae91"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.783475 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-scripts" (OuterVolumeSpecName: "scripts") pod "a9856a93-43ca-4f3d-96a5-50c85086ae91" (UID: "a9856a93-43ca-4f3d-96a5-50c85086ae91"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.783858 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.784134 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.793159 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9856a93-43ca-4f3d-96a5-50c85086ae91-kube-api-access-qq528" (OuterVolumeSpecName: "kube-api-access-qq528") pod "a9856a93-43ca-4f3d-96a5-50c85086ae91" (UID: "a9856a93-43ca-4f3d-96a5-50c85086ae91"). InnerVolumeSpecName "kube-api-access-qq528". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.807834 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a9856a93-43ca-4f3d-96a5-50c85086ae91" (UID: "a9856a93-43ca-4f3d-96a5-50c85086ae91"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.828885 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-f87d658c4-wf95n"] Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.878192 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-config-data\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.878454 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-public-tls-certs\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.878488 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-combined-ca-bundle\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.878519 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-logs\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.878549 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-internal-tls-certs\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.878574 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rx4pb\" (UniqueName: \"kubernetes.io/projected/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-kube-api-access-rx4pb\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.878648 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-config-data-custom\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.878971 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.878997 4578 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.879007 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qq528\" (UniqueName: \"kubernetes.io/projected/a9856a93-43ca-4f3d-96a5-50c85086ae91-kube-api-access-qq528\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.879021 4578 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.891533 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9856a93-43ca-4f3d-96a5-50c85086ae91" (UID: "a9856a93-43ca-4f3d-96a5-50c85086ae91"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.892866 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-config-data" (OuterVolumeSpecName: "config-data") pod "a9856a93-43ca-4f3d-96a5-50c85086ae91" (UID: "a9856a93-43ca-4f3d-96a5-50c85086ae91"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.919335 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.975020 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" event={"ID":"3579029a-5cd2-43d7-9611-7e77a76acd5c","Type":"ContainerDied","Data":"ebd9e191f588b9275e570c4edd5552be6859879c0043efe903f921a6400718c7"} Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.975073 4578 scope.go:117] "RemoveContainer" containerID="5e3db199e76de0240c635a99e2ecc1c9e01f4c85de1237bb45f06ad520a825eb" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.975194 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.981365 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-dns-svc\") pod \"3579029a-5cd2-43d7-9611-7e77a76acd5c\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.981456 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-ovsdbserver-sb\") pod \"3579029a-5cd2-43d7-9611-7e77a76acd5c\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.993224 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-config\") pod \"3579029a-5cd2-43d7-9611-7e77a76acd5c\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.993557 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-dns-swift-storage-0\") pod \"3579029a-5cd2-43d7-9611-7e77a76acd5c\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.993714 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-698hz\" (UniqueName: \"kubernetes.io/projected/3579029a-5cd2-43d7-9611-7e77a76acd5c-kube-api-access-698hz\") pod \"3579029a-5cd2-43d7-9611-7e77a76acd5c\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.993799 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-ovsdbserver-nb\") pod \"3579029a-5cd2-43d7-9611-7e77a76acd5c\" (UID: \"3579029a-5cd2-43d7-9611-7e77a76acd5c\") " Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.994174 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-config-data\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.994310 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-public-tls-certs\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.994417 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-combined-ca-bundle\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.994510 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-logs\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.994586 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-internal-tls-certs\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.994685 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rx4pb\" (UniqueName: \"kubernetes.io/projected/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-kube-api-access-rx4pb\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.994825 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-config-data-custom\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.995007 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:12:59 crc kubenswrapper[4578]: I1003 13:12:59.995071 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9856a93-43ca-4f3d-96a5-50c85086ae91-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.006467 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-logs\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.030920 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jrdpv" event={"ID":"a9856a93-43ca-4f3d-96a5-50c85086ae91","Type":"ContainerDied","Data":"e6518571c6e27ae02edc2365e1598cfd78d3140aca53fac6281ace75cb6eb008"} Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.031154 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6518571c6e27ae02edc2365e1598cfd78d3140aca53fac6281ace75cb6eb008" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.031307 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jrdpv" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.032286 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3579029a-5cd2-43d7-9611-7e77a76acd5c-kube-api-access-698hz" (OuterVolumeSpecName: "kube-api-access-698hz") pod "3579029a-5cd2-43d7-9611-7e77a76acd5c" (UID: "3579029a-5cd2-43d7-9611-7e77a76acd5c"). InnerVolumeSpecName "kube-api-access-698hz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.050350 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-config-data\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.097077 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-internal-tls-certs\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.102659 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-combined-ca-bundle\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.102868 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-config-data-custom\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.103825 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-698hz\" (UniqueName: \"kubernetes.io/projected/3579029a-5cd2-43d7-9611-7e77a76acd5c-kube-api-access-698hz\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.116642 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-public-tls-certs\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.127311 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rx4pb\" (UniqueName: \"kubernetes.io/projected/8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922-kube-api-access-rx4pb\") pod \"barbican-api-f87d658c4-wf95n\" (UID: \"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922\") " pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.188115 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3579029a-5cd2-43d7-9611-7e77a76acd5c" (UID: "3579029a-5cd2-43d7-9611-7e77a76acd5c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.198554 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.214228 4578 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.215084 4578 scope.go:117] "RemoveContainer" containerID="4239f81a6442d31df07d455e6472259d868ed6c5074d66df49a2660b6459b68d" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.228851 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3579029a-5cd2-43d7-9611-7e77a76acd5c" (UID: "3579029a-5cd2-43d7-9611-7e77a76acd5c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.251063 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-config" (OuterVolumeSpecName: "config") pod "3579029a-5cd2-43d7-9611-7e77a76acd5c" (UID: "3579029a-5cd2-43d7-9611-7e77a76acd5c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.255735 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3579029a-5cd2-43d7-9611-7e77a76acd5c" (UID: "3579029a-5cd2-43d7-9611-7e77a76acd5c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.275852 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3579029a-5cd2-43d7-9611-7e77a76acd5c" (UID: "3579029a-5cd2-43d7-9611-7e77a76acd5c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.322258 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.322294 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.322308 4578 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.322318 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3579029a-5cd2-43d7-9611-7e77a76acd5c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.418408 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-2gwvm"] Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.458732 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-2gwvm"] Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.876966 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-64b9f57cc5-jrm9t"] Oct 03 13:13:00 crc kubenswrapper[4578]: E1003 13:13:00.877601 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3579029a-5cd2-43d7-9611-7e77a76acd5c" containerName="dnsmasq-dns" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.877612 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="3579029a-5cd2-43d7-9611-7e77a76acd5c" containerName="dnsmasq-dns" Oct 03 13:13:00 crc kubenswrapper[4578]: E1003 13:13:00.877641 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3579029a-5cd2-43d7-9611-7e77a76acd5c" containerName="init" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.877647 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="3579029a-5cd2-43d7-9611-7e77a76acd5c" containerName="init" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.887379 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="3579029a-5cd2-43d7-9611-7e77a76acd5c" containerName="dnsmasq-dns" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.892737 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.898247 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.898453 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wxw7h" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.898560 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.898726 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.898781 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.898724 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 03 13:13:00 crc kubenswrapper[4578]: I1003 13:13:00.902516 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-64b9f57cc5-jrm9t"] Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.058808 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-fernet-keys\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.058866 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-internal-tls-certs\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.058913 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-credential-keys\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.058939 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-scripts\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.058967 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-combined-ca-bundle\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.059003 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbth2\" (UniqueName: \"kubernetes.io/projected/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-kube-api-access-tbth2\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.059044 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-config-data\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.059081 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-public-tls-certs\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.133187 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3579029a-5cd2-43d7-9611-7e77a76acd5c" path="/var/lib/kubelet/pods/3579029a-5cd2-43d7-9611-7e77a76acd5c/volumes" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.135696 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-786f64db8b-9n26w"] Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.135734 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-d9f65c5d7-qf4l9"] Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.143213 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6228c849-5824-4a21-8318-754c65d79a2e","Type":"ContainerStarted","Data":"ed71ed5afef9b00e0ea00c8b27a6f35651ee63ca3bdc776cb10be4964a07b008"} Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.146453 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-d9f65c5d7-qf4l9" event={"ID":"c187292d-a86d-455d-9c52-dabc5c3b4631","Type":"ContainerStarted","Data":"e245384396561637c9133444dab48ce8e6f70a18863a6ea34371f9e5758e9c64"} Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.160544 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-internal-tls-certs\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.160599 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-credential-keys\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.160618 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-scripts\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.160649 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-combined-ca-bundle\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.160675 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbth2\" (UniqueName: \"kubernetes.io/projected/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-kube-api-access-tbth2\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.160732 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-config-data\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.160762 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-public-tls-certs\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.160843 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-fernet-keys\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.189123 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-fernet-keys\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.215197 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-credential-keys\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.231278 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-scripts\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.235672 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-config-data\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.246192 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-public-tls-certs\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.251166 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-combined-ca-bundle\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.255830 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-internal-tls-certs\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.288073 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbth2\" (UniqueName: \"kubernetes.io/projected/f70c10e6-f3e1-41e0-a30f-267ca3a05c82-kube-api-access-tbth2\") pod \"keystone-64b9f57cc5-jrm9t\" (UID: \"f70c10e6-f3e1-41e0-a30f-267ca3a05c82\") " pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.316086 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-66cbcb46d8-9jksg"] Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.343966 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.700396 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-f87d658c4-wf95n"] Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.766435 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-x5hxs"] Oct 03 13:13:01 crc kubenswrapper[4578]: I1003 13:13:01.823512 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-85d5675966-rb6nf"] Oct 03 13:13:01 crc kubenswrapper[4578]: W1003 13:13:01.873095 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb40b382f_7fb7_4f67_9237_e5148abf7610.slice/crio-e37342f43e0cb2ae68aa5a669f4a4fe93deff06fcbc18c195742a65adb06a46b WatchSource:0}: Error finding container e37342f43e0cb2ae68aa5a669f4a4fe93deff06fcbc18c195742a65adb06a46b: Status 404 returned error can't find the container with id e37342f43e0cb2ae68aa5a669f4a4fe93deff06fcbc18c195742a65adb06a46b Oct 03 13:13:02 crc kubenswrapper[4578]: I1003 13:13:02.217683 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f87d658c4-wf95n" event={"ID":"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922","Type":"ContainerStarted","Data":"c3191b142a50d1c994080656139d53360230bcae63a5ff89f309831ded86a548"} Oct 03 13:13:02 crc kubenswrapper[4578]: I1003 13:13:02.241747 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-66cbcb46d8-9jksg" event={"ID":"5e89457e-7c3d-4f85-a14d-ca47f5bdca07","Type":"ContainerStarted","Data":"3208bd34a26796732b649b40b3c6d0482c15e802ab6ed8c83a0161ffa7ba47ba"} Oct 03 13:13:02 crc kubenswrapper[4578]: I1003 13:13:02.241789 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-66cbcb46d8-9jksg" event={"ID":"5e89457e-7c3d-4f85-a14d-ca47f5bdca07","Type":"ContainerStarted","Data":"d1590a33308f9abc24cfbeb7336fbc26e44611980ef7a0dac3b73e4c442b4480"} Oct 03 13:13:02 crc kubenswrapper[4578]: I1003 13:13:02.274949 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-q6ppj" event={"ID":"6b310034-f02f-48ac-a47a-4f41f15379a5","Type":"ContainerStarted","Data":"4a7daca41c304b21cfd34068e7c4f88933b9c6c3148f24319a72504cd0d21be2"} Oct 03 13:13:02 crc kubenswrapper[4578]: I1003 13:13:02.319385 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"25ef994a-0092-4e72-a636-f51bbd1a8a99","Type":"ContainerStarted","Data":"2323557f12cf7f4bfe4a74ca5af7e4ce03b61bb618e398e5eba8cfa1c269bcda"} Oct 03 13:13:02 crc kubenswrapper[4578]: I1003 13:13:02.320354 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-q6ppj" podStartSLOduration=4.776217589 podStartE2EDuration="56.320337048s" podCreationTimestamp="2025-10-03 13:12:06 +0000 UTC" firstStartedPulling="2025-10-03 13:12:08.399077162 +0000 UTC m=+1264.197549346" lastFinishedPulling="2025-10-03 13:12:59.943196621 +0000 UTC m=+1315.741668805" observedRunningTime="2025-10-03 13:13:02.302913675 +0000 UTC m=+1318.101385859" watchObservedRunningTime="2025-10-03 13:13:02.320337048 +0000 UTC m=+1318.118809232" Oct 03 13:13:02 crc kubenswrapper[4578]: I1003 13:13:02.348048 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" event={"ID":"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc","Type":"ContainerStarted","Data":"d23ffb0337c0372d2785ff2edb883d34fc8ae4b387819c47f2173f0d34ec780f"} Oct 03 13:13:02 crc kubenswrapper[4578]: I1003 13:13:02.361147 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-64b9f57cc5-jrm9t"] Oct 03 13:13:02 crc kubenswrapper[4578]: I1003 13:13:02.362112 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=15.362080949 podStartE2EDuration="15.362080949s" podCreationTimestamp="2025-10-03 13:12:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:13:02.356302019 +0000 UTC m=+1318.154774203" watchObservedRunningTime="2025-10-03 13:13:02.362080949 +0000 UTC m=+1318.160553133" Oct 03 13:13:02 crc kubenswrapper[4578]: I1003 13:13:02.367914 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85d5675966-rb6nf" event={"ID":"b40b382f-7fb7-4f67-9237-e5148abf7610","Type":"ContainerStarted","Data":"e37342f43e0cb2ae68aa5a669f4a4fe93deff06fcbc18c195742a65adb06a46b"} Oct 03 13:13:02 crc kubenswrapper[4578]: I1003 13:13:02.380741 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" event={"ID":"3bb09a8d-f51b-494d-8a75-8690e842a916","Type":"ContainerStarted","Data":"6842712bafc9dfe51cfbaaca22315052823cd810156f48aad8f52a7c04eac3a9"} Oct 03 13:13:02 crc kubenswrapper[4578]: I1003 13:13:02.760716 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8b5c85b87-2gwvm" podUID="3579029a-5cd2-43d7-9611-7e77a76acd5c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: i/o timeout" Oct 03 13:13:03 crc kubenswrapper[4578]: I1003 13:13:03.410567 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-66cbcb46d8-9jksg" event={"ID":"5e89457e-7c3d-4f85-a14d-ca47f5bdca07","Type":"ContainerStarted","Data":"a6b935a3663cdf097bd91114a3326c8bee92070ed06d19c1deea632064f9cb3c"} Oct 03 13:13:03 crc kubenswrapper[4578]: I1003 13:13:03.413074 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-64b9f57cc5-jrm9t" event={"ID":"f70c10e6-f3e1-41e0-a30f-267ca3a05c82","Type":"ContainerStarted","Data":"22e22b38bdd7bc044aa09ef9dfd503cd6a026d13745cc8c7b0424c04dd9a6471"} Oct 03 13:13:03 crc kubenswrapper[4578]: I1003 13:13:03.413114 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-64b9f57cc5-jrm9t" event={"ID":"f70c10e6-f3e1-41e0-a30f-267ca3a05c82","Type":"ContainerStarted","Data":"c9f83e38019fd4ceb04ad94593fba0f947d4ff9eb37ffeac6ad1040004b72368"} Oct 03 13:13:03 crc kubenswrapper[4578]: I1003 13:13:03.413504 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:03 crc kubenswrapper[4578]: I1003 13:13:03.416879 4578 generic.go:334] "Generic (PLEG): container finished" podID="9fdb8751-6b7b-4a75-abe6-0d5d5428cccc" containerID="16cee472c5d94cf61cd4e5b6a1916995a6c585a2b77b67fcf84c9b254b384d44" exitCode=0 Oct 03 13:13:03 crc kubenswrapper[4578]: I1003 13:13:03.416932 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" event={"ID":"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc","Type":"ContainerDied","Data":"16cee472c5d94cf61cd4e5b6a1916995a6c585a2b77b67fcf84c9b254b384d44"} Oct 03 13:13:03 crc kubenswrapper[4578]: I1003 13:13:03.419219 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85d5675966-rb6nf" event={"ID":"b40b382f-7fb7-4f67-9237-e5148abf7610","Type":"ContainerStarted","Data":"9bfdb147e2aee44432322ce8075c6137266854d584755636f7e42e56c26ab192"} Oct 03 13:13:03 crc kubenswrapper[4578]: I1003 13:13:03.419244 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85d5675966-rb6nf" event={"ID":"b40b382f-7fb7-4f67-9237-e5148abf7610","Type":"ContainerStarted","Data":"61dc6fff53873a9950a28c8817606b415df204b784b5fccbc3d21327dc5f282d"} Oct 03 13:13:03 crc kubenswrapper[4578]: I1003 13:13:03.419728 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:13:03 crc kubenswrapper[4578]: I1003 13:13:03.419753 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:13:03 crc kubenswrapper[4578]: I1003 13:13:03.434766 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-66cbcb46d8-9jksg" podStartSLOduration=13.434752576 podStartE2EDuration="13.434752576s" podCreationTimestamp="2025-10-03 13:12:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:13:03.433060794 +0000 UTC m=+1319.231532978" watchObservedRunningTime="2025-10-03 13:13:03.434752576 +0000 UTC m=+1319.233224760" Oct 03 13:13:03 crc kubenswrapper[4578]: I1003 13:13:03.471426 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f87d658c4-wf95n" event={"ID":"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922","Type":"ContainerStarted","Data":"96cf3ea853e42d846ceeba45d60def4d0d9581bead7a75427faa2267c86bd51b"} Oct 03 13:13:03 crc kubenswrapper[4578]: I1003 13:13:03.471476 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f87d658c4-wf95n" event={"ID":"8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922","Type":"ContainerStarted","Data":"b73bdf97b7285089ce842df29a6d93e0ef3977ae51ff198eec57e45478b2ea0d"} Oct 03 13:13:03 crc kubenswrapper[4578]: I1003 13:13:03.471519 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:13:03 crc kubenswrapper[4578]: I1003 13:13:03.471988 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:13:03 crc kubenswrapper[4578]: I1003 13:13:03.558256 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-64b9f57cc5-jrm9t" podStartSLOduration=3.558235954 podStartE2EDuration="3.558235954s" podCreationTimestamp="2025-10-03 13:13:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:13:03.523657947 +0000 UTC m=+1319.322130151" watchObservedRunningTime="2025-10-03 13:13:03.558235954 +0000 UTC m=+1319.356708138" Oct 03 13:13:03 crc kubenswrapper[4578]: I1003 13:13:03.579950 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-85d5675966-rb6nf" podStartSLOduration=7.57992683 podStartE2EDuration="7.57992683s" podCreationTimestamp="2025-10-03 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:13:03.540986246 +0000 UTC m=+1319.339458440" watchObservedRunningTime="2025-10-03 13:13:03.57992683 +0000 UTC m=+1319.378399014" Oct 03 13:13:03 crc kubenswrapper[4578]: I1003 13:13:03.629972 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-f87d658c4-wf95n" podStartSLOduration=4.629951228 podStartE2EDuration="4.629951228s" podCreationTimestamp="2025-10-03 13:12:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:13:03.586499625 +0000 UTC m=+1319.384971819" watchObservedRunningTime="2025-10-03 13:13:03.629951228 +0000 UTC m=+1319.428423412" Oct 03 13:13:04 crc kubenswrapper[4578]: I1003 13:13:04.477676 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" event={"ID":"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc","Type":"ContainerStarted","Data":"ef81916cf4cbd206da14a37bca646adb216da11a3d32cc1f3251c3549ad12503"} Oct 03 13:13:04 crc kubenswrapper[4578]: I1003 13:13:04.478244 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:13:04 crc kubenswrapper[4578]: I1003 13:13:04.478300 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:13:04 crc kubenswrapper[4578]: I1003 13:13:04.478310 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:13:04 crc kubenswrapper[4578]: I1003 13:13:04.503307 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" podStartSLOduration=8.503285374 podStartE2EDuration="8.503285374s" podCreationTimestamp="2025-10-03 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:13:04.49577691 +0000 UTC m=+1320.294249094" watchObservedRunningTime="2025-10-03 13:13:04.503285374 +0000 UTC m=+1320.301757558" Oct 03 13:13:06 crc kubenswrapper[4578]: I1003 13:13:06.781016 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7b58cf9568-f8nng" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 03 13:13:06 crc kubenswrapper[4578]: I1003 13:13:06.948461 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-68684899bd-2cwll" podUID="d49d623e-2cd6-4786-b1da-34b78b89a134" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 03 13:13:07 crc kubenswrapper[4578]: I1003 13:13:07.372000 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 03 13:13:07 crc kubenswrapper[4578]: I1003 13:13:07.372130 4578 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 13:13:07 crc kubenswrapper[4578]: I1003 13:13:07.529771 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 03 13:13:07 crc kubenswrapper[4578]: I1003 13:13:07.545345 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" event={"ID":"3bb09a8d-f51b-494d-8a75-8690e842a916","Type":"ContainerStarted","Data":"a91d4987c29439aa6843aa008a0f5645b5a63825630998be39ac834ef0f95e3a"} Oct 03 13:13:07 crc kubenswrapper[4578]: I1003 13:13:07.545393 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" event={"ID":"3bb09a8d-f51b-494d-8a75-8690e842a916","Type":"ContainerStarted","Data":"1384edc1b7090b212b17cfcbf35dc2b989fc1f4b0f156ef0cc07146fa9f13969"} Oct 03 13:13:07 crc kubenswrapper[4578]: I1003 13:13:07.574687 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-d9f65c5d7-qf4l9" event={"ID":"c187292d-a86d-455d-9c52-dabc5c3b4631","Type":"ContainerStarted","Data":"50b5f680c8dc27d0fe610c34932f3a486ac80eeebb4c8ea7386368fe54fd1e48"} Oct 03 13:13:07 crc kubenswrapper[4578]: I1003 13:13:07.574734 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-d9f65c5d7-qf4l9" event={"ID":"c187292d-a86d-455d-9c52-dabc5c3b4631","Type":"ContainerStarted","Data":"03d1d2b79ce9c030201b364061702e534640cb2ca49f4ed146695d4a055d745b"} Oct 03 13:13:07 crc kubenswrapper[4578]: I1003 13:13:07.577347 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-786f64db8b-9n26w" podStartSLOduration=7.6107161869999995 podStartE2EDuration="12.577330678s" podCreationTimestamp="2025-10-03 13:12:55 +0000 UTC" firstStartedPulling="2025-10-03 13:13:01.145781226 +0000 UTC m=+1316.944253410" lastFinishedPulling="2025-10-03 13:13:06.112395717 +0000 UTC m=+1321.910867901" observedRunningTime="2025-10-03 13:13:07.565700705 +0000 UTC m=+1323.364172899" watchObservedRunningTime="2025-10-03 13:13:07.577330678 +0000 UTC m=+1323.375802862" Oct 03 13:13:08 crc kubenswrapper[4578]: I1003 13:13:08.652771 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 03 13:13:08 crc kubenswrapper[4578]: I1003 13:13:08.652829 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 03 13:13:08 crc kubenswrapper[4578]: I1003 13:13:08.688291 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 03 13:13:08 crc kubenswrapper[4578]: I1003 13:13:08.702309 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 03 13:13:08 crc kubenswrapper[4578]: I1003 13:13:08.718086 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-d9f65c5d7-qf4l9" podStartSLOduration=8.643803991 podStartE2EDuration="13.718064996s" podCreationTimestamp="2025-10-03 13:12:55 +0000 UTC" firstStartedPulling="2025-10-03 13:13:01.143079712 +0000 UTC m=+1316.941551896" lastFinishedPulling="2025-10-03 13:13:06.217340717 +0000 UTC m=+1322.015812901" observedRunningTime="2025-10-03 13:13:07.611836133 +0000 UTC m=+1323.410308327" watchObservedRunningTime="2025-10-03 13:13:08.718064996 +0000 UTC m=+1324.516537180" Oct 03 13:13:09 crc kubenswrapper[4578]: I1003 13:13:09.619232 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 03 13:13:09 crc kubenswrapper[4578]: I1003 13:13:09.619564 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 03 13:13:11 crc kubenswrapper[4578]: I1003 13:13:11.637240 4578 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 13:13:11 crc kubenswrapper[4578]: I1003 13:13:11.637265 4578 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 13:13:11 crc kubenswrapper[4578]: I1003 13:13:11.646806 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:13:11 crc kubenswrapper[4578]: I1003 13:13:11.711368 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-phtfx"] Oct 03 13:13:11 crc kubenswrapper[4578]: I1003 13:13:11.711590 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" podUID="b85d10a6-888d-4862-986f-5e44b8d3d229" containerName="dnsmasq-dns" containerID="cri-o://460acd0e4b7cdbdda5a53c34a461e8bf587fbb22d1ee478ae0ec8f7d216a1abc" gracePeriod=10 Oct 03 13:13:12 crc kubenswrapper[4578]: I1003 13:13:12.501665 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:13:12 crc kubenswrapper[4578]: I1003 13:13:12.660368 4578 generic.go:334] "Generic (PLEG): container finished" podID="b85d10a6-888d-4862-986f-5e44b8d3d229" containerID="460acd0e4b7cdbdda5a53c34a461e8bf587fbb22d1ee478ae0ec8f7d216a1abc" exitCode=0 Oct 03 13:13:12 crc kubenswrapper[4578]: I1003 13:13:12.660427 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" event={"ID":"b85d10a6-888d-4862-986f-5e44b8d3d229","Type":"ContainerDied","Data":"460acd0e4b7cdbdda5a53c34a461e8bf587fbb22d1ee478ae0ec8f7d216a1abc"} Oct 03 13:13:13 crc kubenswrapper[4578]: I1003 13:13:13.682076 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c684ff4d9-2h9cq" event={"ID":"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b","Type":"ContainerDied","Data":"80929f7ff9f2db6421b73861631c698d81ca2c15a2ea97522dc9d1eb510196b2"} Oct 03 13:13:13 crc kubenswrapper[4578]: I1003 13:13:13.682034 4578 generic.go:334] "Generic (PLEG): container finished" podID="1b0f9e3d-a619-437f-9339-d61b6ffb8d4b" containerID="80929f7ff9f2db6421b73861631c698d81ca2c15a2ea97522dc9d1eb510196b2" exitCode=137 Oct 03 13:13:13 crc kubenswrapper[4578]: I1003 13:13:13.682878 4578 generic.go:334] "Generic (PLEG): container finished" podID="1b0f9e3d-a619-437f-9339-d61b6ffb8d4b" containerID="5402c4c9ddc18e14fe296ec7864a7e4a97e4da6535530acc67d68775074d1046" exitCode=137 Oct 03 13:13:13 crc kubenswrapper[4578]: I1003 13:13:13.682903 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c684ff4d9-2h9cq" event={"ID":"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b","Type":"ContainerDied","Data":"5402c4c9ddc18e14fe296ec7864a7e4a97e4da6535530acc67d68775074d1046"} Oct 03 13:13:13 crc kubenswrapper[4578]: I1003 13:13:13.942483 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-85d5675966-rb6nf" podUID="b40b382f-7fb7-4f67-9237-e5148abf7610" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 13:13:13 crc kubenswrapper[4578]: I1003 13:13:13.942917 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-85d5675966-rb6nf" podUID="b40b382f-7fb7-4f67-9237-e5148abf7610" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 13:13:14 crc kubenswrapper[4578]: I1003 13:13:14.215869 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-f87d658c4-wf95n" podUID="8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:13:14 crc kubenswrapper[4578]: I1003 13:13:14.216571 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-f87d658c4-wf95n" podUID="8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:13:14 crc kubenswrapper[4578]: I1003 13:13:14.508155 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 03 13:13:14 crc kubenswrapper[4578]: I1003 13:13:14.508274 4578 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 13:13:14 crc kubenswrapper[4578]: I1003 13:13:14.526601 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:13:14 crc kubenswrapper[4578]: I1003 13:13:14.541975 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:13:14 crc kubenswrapper[4578]: I1003 13:13:14.702196 4578 generic.go:334] "Generic (PLEG): container finished" podID="6b310034-f02f-48ac-a47a-4f41f15379a5" containerID="4a7daca41c304b21cfd34068e7c4f88933b9c6c3148f24319a72504cd0d21be2" exitCode=0 Oct 03 13:13:14 crc kubenswrapper[4578]: I1003 13:13:14.702254 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-q6ppj" event={"ID":"6b310034-f02f-48ac-a47a-4f41f15379a5","Type":"ContainerDied","Data":"4a7daca41c304b21cfd34068e7c4f88933b9c6c3148f24319a72504cd0d21be2"} Oct 03 13:13:14 crc kubenswrapper[4578]: I1003 13:13:14.708258 4578 generic.go:334] "Generic (PLEG): container finished" podID="a567645b-5ae1-46a0-86b1-e24b1592e060" containerID="4b3f4f06273a71501b5530c10beab91d989254af38cef506f55056b9e26f0b06" exitCode=137 Oct 03 13:13:14 crc kubenswrapper[4578]: I1003 13:13:14.708284 4578 generic.go:334] "Generic (PLEG): container finished" podID="a567645b-5ae1-46a0-86b1-e24b1592e060" containerID="00d4bed9ee5fe90eebe7ce2ec9be50c6902436a20863280d15aafd3f0158f87d" exitCode=137 Oct 03 13:13:14 crc kubenswrapper[4578]: I1003 13:13:14.708302 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d5f486769-rxsds" event={"ID":"a567645b-5ae1-46a0-86b1-e24b1592e060","Type":"ContainerDied","Data":"4b3f4f06273a71501b5530c10beab91d989254af38cef506f55056b9e26f0b06"} Oct 03 13:13:14 crc kubenswrapper[4578]: I1003 13:13:14.708321 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d5f486769-rxsds" event={"ID":"a567645b-5ae1-46a0-86b1-e24b1592e060","Type":"ContainerDied","Data":"00d4bed9ee5fe90eebe7ce2ec9be50c6902436a20863280d15aafd3f0158f87d"} Oct 03 13:13:15 crc kubenswrapper[4578]: I1003 13:13:15.029569 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:13:15 crc kubenswrapper[4578]: I1003 13:13:15.115129 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 03 13:13:15 crc kubenswrapper[4578]: I1003 13:13:15.159509 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-d7d5c9485-l2ww5" Oct 03 13:13:15 crc kubenswrapper[4578]: I1003 13:13:15.209057 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-f87d658c4-wf95n" podUID="8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:13:15 crc kubenswrapper[4578]: I1003 13:13:15.253924 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6d67444cbd-9qgkn"] Oct 03 13:13:15 crc kubenswrapper[4578]: I1003 13:13:15.254125 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6d67444cbd-9qgkn" podUID="101089c9-6f91-42fb-a5ab-d907c80ff069" containerName="neutron-api" containerID="cri-o://e73cf2c70f9bc072fedc51586fa3a7d7b4c59719b48f280707a595fd70f5d102" gracePeriod=30 Oct 03 13:13:15 crc kubenswrapper[4578]: I1003 13:13:15.254718 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6d67444cbd-9qgkn" podUID="101089c9-6f91-42fb-a5ab-d907c80ff069" containerName="neutron-httpd" containerID="cri-o://6617906f8965d5b3d283bc14710803835bf60a5cb3e056a9c10f0e336b4cdff7" gracePeriod=30 Oct 03 13:13:16 crc kubenswrapper[4578]: I1003 13:13:16.243800 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" podUID="b85d10a6-888d-4862-986f-5e44b8d3d229" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.149:5353: connect: connection refused" Oct 03 13:13:16 crc kubenswrapper[4578]: I1003 13:13:16.748485 4578 generic.go:334] "Generic (PLEG): container finished" podID="101089c9-6f91-42fb-a5ab-d907c80ff069" containerID="6617906f8965d5b3d283bc14710803835bf60a5cb3e056a9c10f0e336b4cdff7" exitCode=0 Oct 03 13:13:16 crc kubenswrapper[4578]: I1003 13:13:16.748790 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6d67444cbd-9qgkn" event={"ID":"101089c9-6f91-42fb-a5ab-d907c80ff069","Type":"ContainerDied","Data":"6617906f8965d5b3d283bc14710803835bf60a5cb3e056a9c10f0e336b4cdff7"} Oct 03 13:13:16 crc kubenswrapper[4578]: I1003 13:13:16.781141 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7b58cf9568-f8nng" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 03 13:13:16 crc kubenswrapper[4578]: I1003 13:13:16.781205 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:13:16 crc kubenswrapper[4578]: I1003 13:13:16.781937 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"033a127d7b3dc6a5469b25fc2682629ad13d3e1ed8dd904aad8edc6112ea940b"} pod="openstack/horizon-7b58cf9568-f8nng" containerMessage="Container horizon failed startup probe, will be restarted" Oct 03 13:13:16 crc kubenswrapper[4578]: I1003 13:13:16.781966 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7b58cf9568-f8nng" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerName="horizon" containerID="cri-o://033a127d7b3dc6a5469b25fc2682629ad13d3e1ed8dd904aad8edc6112ea940b" gracePeriod=30 Oct 03 13:13:16 crc kubenswrapper[4578]: I1003 13:13:16.948605 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-68684899bd-2cwll" podUID="d49d623e-2cd6-4786-b1da-34b78b89a134" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 03 13:13:16 crc kubenswrapper[4578]: I1003 13:13:16.948744 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:13:16 crc kubenswrapper[4578]: I1003 13:13:16.949458 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"043ba4870e4db2272214244e538587f588862b6e71b023c4ea7dfa6638da57bd"} pod="openstack/horizon-68684899bd-2cwll" containerMessage="Container horizon failed startup probe, will be restarted" Oct 03 13:13:16 crc kubenswrapper[4578]: I1003 13:13:16.949493 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-68684899bd-2cwll" podUID="d49d623e-2cd6-4786-b1da-34b78b89a134" containerName="horizon" containerID="cri-o://043ba4870e4db2272214244e538587f588862b6e71b023c4ea7dfa6638da57bd" gracePeriod=30 Oct 03 13:13:17 crc kubenswrapper[4578]: I1003 13:13:17.784098 4578 generic.go:334] "Generic (PLEG): container finished" podID="101089c9-6f91-42fb-a5ab-d907c80ff069" containerID="e73cf2c70f9bc072fedc51586fa3a7d7b4c59719b48f280707a595fd70f5d102" exitCode=0 Oct 03 13:13:17 crc kubenswrapper[4578]: I1003 13:13:17.784814 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6d67444cbd-9qgkn" event={"ID":"101089c9-6f91-42fb-a5ab-d907c80ff069","Type":"ContainerDied","Data":"e73cf2c70f9bc072fedc51586fa3a7d7b4c59719b48f280707a595fd70f5d102"} Oct 03 13:13:18 crc kubenswrapper[4578]: I1003 13:13:18.561201 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-f87d658c4-wf95n" Oct 03 13:13:18 crc kubenswrapper[4578]: I1003 13:13:18.635090 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-85d5675966-rb6nf"] Oct 03 13:13:18 crc kubenswrapper[4578]: I1003 13:13:18.635387 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-85d5675966-rb6nf" podUID="b40b382f-7fb7-4f67-9237-e5148abf7610" containerName="barbican-api-log" containerID="cri-o://61dc6fff53873a9950a28c8817606b415df204b784b5fccbc3d21327dc5f282d" gracePeriod=30 Oct 03 13:13:18 crc kubenswrapper[4578]: I1003 13:13:18.635710 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-85d5675966-rb6nf" podUID="b40b382f-7fb7-4f67-9237-e5148abf7610" containerName="barbican-api" containerID="cri-o://9bfdb147e2aee44432322ce8075c6137266854d584755636f7e42e56c26ab192" gracePeriod=30 Oct 03 13:13:18 crc kubenswrapper[4578]: I1003 13:13:18.672866 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-85d5675966-rb6nf" podUID="b40b382f-7fb7-4f67-9237-e5148abf7610" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": EOF" Oct 03 13:13:18 crc kubenswrapper[4578]: I1003 13:13:18.672930 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-85d5675966-rb6nf" podUID="b40b382f-7fb7-4f67-9237-e5148abf7610" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": EOF" Oct 03 13:13:18 crc kubenswrapper[4578]: I1003 13:13:18.673003 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-85d5675966-rb6nf" podUID="b40b382f-7fb7-4f67-9237-e5148abf7610" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": EOF" Oct 03 13:13:18 crc kubenswrapper[4578]: I1003 13:13:18.673069 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-85d5675966-rb6nf" podUID="b40b382f-7fb7-4f67-9237-e5148abf7610" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": EOF" Oct 03 13:13:18 crc kubenswrapper[4578]: I1003 13:13:18.710818 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-85d5675966-rb6nf" podUID="b40b382f-7fb7-4f67-9237-e5148abf7610" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": EOF" Oct 03 13:13:18 crc kubenswrapper[4578]: I1003 13:13:18.808198 4578 generic.go:334] "Generic (PLEG): container finished" podID="b40b382f-7fb7-4f67-9237-e5148abf7610" containerID="61dc6fff53873a9950a28c8817606b415df204b784b5fccbc3d21327dc5f282d" exitCode=143 Oct 03 13:13:18 crc kubenswrapper[4578]: I1003 13:13:18.808259 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85d5675966-rb6nf" event={"ID":"b40b382f-7fb7-4f67-9237-e5148abf7610","Type":"ContainerDied","Data":"61dc6fff53873a9950a28c8817606b415df204b784b5fccbc3d21327dc5f282d"} Oct 03 13:13:20 crc kubenswrapper[4578]: E1003 13:13:20.809412 4578 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Oct 03 13:13:20 crc kubenswrapper[4578]: E1003 13:13:20.810070 4578 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tbm4c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(6228c849-5824-4a21-8318-754c65d79a2e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 03 13:13:20 crc kubenswrapper[4578]: E1003 13:13:20.811532 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="6228c849-5824-4a21-8318-754c65d79a2e" Oct 03 13:13:20 crc kubenswrapper[4578]: I1003 13:13:20.833304 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-q6ppj" event={"ID":"6b310034-f02f-48ac-a47a-4f41f15379a5","Type":"ContainerDied","Data":"f4056975c2c754b545287c210946a5435f841c2b8b9c3ebb99ad66237cb15481"} Oct 03 13:13:20 crc kubenswrapper[4578]: I1003 13:13:20.833369 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4056975c2c754b545287c210946a5435f841c2b8b9c3ebb99ad66237cb15481" Oct 03 13:13:20 crc kubenswrapper[4578]: I1003 13:13:20.833393 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6228c849-5824-4a21-8318-754c65d79a2e" containerName="ceilometer-notification-agent" containerID="cri-o://b45d39e61e8fc0942f781f361e7f58c723585ade153a008212d11232868eaf4d" gracePeriod=30 Oct 03 13:13:20 crc kubenswrapper[4578]: I1003 13:13:20.833421 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6228c849-5824-4a21-8318-754c65d79a2e" containerName="sg-core" containerID="cri-o://ed71ed5afef9b00e0ea00c8b27a6f35651ee63ca3bdc776cb10be4964a07b008" gracePeriod=30 Oct 03 13:13:20 crc kubenswrapper[4578]: I1003 13:13:20.913150 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.015370 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55whk\" (UniqueName: \"kubernetes.io/projected/6b310034-f02f-48ac-a47a-4f41f15379a5-kube-api-access-55whk\") pod \"6b310034-f02f-48ac-a47a-4f41f15379a5\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.015473 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-scripts\") pod \"6b310034-f02f-48ac-a47a-4f41f15379a5\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.015514 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-config-data\") pod \"6b310034-f02f-48ac-a47a-4f41f15379a5\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.015557 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b310034-f02f-48ac-a47a-4f41f15379a5-etc-machine-id\") pod \"6b310034-f02f-48ac-a47a-4f41f15379a5\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.015649 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-db-sync-config-data\") pod \"6b310034-f02f-48ac-a47a-4f41f15379a5\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.015690 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-combined-ca-bundle\") pod \"6b310034-f02f-48ac-a47a-4f41f15379a5\" (UID: \"6b310034-f02f-48ac-a47a-4f41f15379a5\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.016850 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6b310034-f02f-48ac-a47a-4f41f15379a5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6b310034-f02f-48ac-a47a-4f41f15379a5" (UID: "6b310034-f02f-48ac-a47a-4f41f15379a5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.023115 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b310034-f02f-48ac-a47a-4f41f15379a5-kube-api-access-55whk" (OuterVolumeSpecName: "kube-api-access-55whk") pod "6b310034-f02f-48ac-a47a-4f41f15379a5" (UID: "6b310034-f02f-48ac-a47a-4f41f15379a5"). InnerVolumeSpecName "kube-api-access-55whk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.024044 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-scripts" (OuterVolumeSpecName: "scripts") pod "6b310034-f02f-48ac-a47a-4f41f15379a5" (UID: "6b310034-f02f-48ac-a47a-4f41f15379a5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.024758 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6b310034-f02f-48ac-a47a-4f41f15379a5" (UID: "6b310034-f02f-48ac-a47a-4f41f15379a5"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.057439 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b310034-f02f-48ac-a47a-4f41f15379a5" (UID: "6b310034-f02f-48ac-a47a-4f41f15379a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.083223 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-config-data" (OuterVolumeSpecName: "config-data") pod "6b310034-f02f-48ac-a47a-4f41f15379a5" (UID: "6b310034-f02f-48ac-a47a-4f41f15379a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.118146 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.118183 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.118196 4578 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6b310034-f02f-48ac-a47a-4f41f15379a5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.118209 4578 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.118220 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b310034-f02f-48ac-a47a-4f41f15379a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.118233 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55whk\" (UniqueName: \"kubernetes.io/projected/6b310034-f02f-48ac-a47a-4f41f15379a5-kube-api-access-55whk\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.245609 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" podUID="b85d10a6-888d-4862-986f-5e44b8d3d229" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.149:5353: connect: connection refused" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.352106 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d5f486769-rxsds" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.423186 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dv5wc\" (UniqueName: \"kubernetes.io/projected/a567645b-5ae1-46a0-86b1-e24b1592e060-kube-api-access-dv5wc\") pod \"a567645b-5ae1-46a0-86b1-e24b1592e060\" (UID: \"a567645b-5ae1-46a0-86b1-e24b1592e060\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.423270 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a567645b-5ae1-46a0-86b1-e24b1592e060-scripts\") pod \"a567645b-5ae1-46a0-86b1-e24b1592e060\" (UID: \"a567645b-5ae1-46a0-86b1-e24b1592e060\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.423348 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a567645b-5ae1-46a0-86b1-e24b1592e060-logs\") pod \"a567645b-5ae1-46a0-86b1-e24b1592e060\" (UID: \"a567645b-5ae1-46a0-86b1-e24b1592e060\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.423406 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a567645b-5ae1-46a0-86b1-e24b1592e060-config-data\") pod \"a567645b-5ae1-46a0-86b1-e24b1592e060\" (UID: \"a567645b-5ae1-46a0-86b1-e24b1592e060\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.423445 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a567645b-5ae1-46a0-86b1-e24b1592e060-horizon-secret-key\") pod \"a567645b-5ae1-46a0-86b1-e24b1592e060\" (UID: \"a567645b-5ae1-46a0-86b1-e24b1592e060\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.427401 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a567645b-5ae1-46a0-86b1-e24b1592e060-logs" (OuterVolumeSpecName: "logs") pod "a567645b-5ae1-46a0-86b1-e24b1592e060" (UID: "a567645b-5ae1-46a0-86b1-e24b1592e060"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.438738 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a567645b-5ae1-46a0-86b1-e24b1592e060-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a567645b-5ae1-46a0-86b1-e24b1592e060" (UID: "a567645b-5ae1-46a0-86b1-e24b1592e060"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.448009 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a567645b-5ae1-46a0-86b1-e24b1592e060-kube-api-access-dv5wc" (OuterVolumeSpecName: "kube-api-access-dv5wc") pod "a567645b-5ae1-46a0-86b1-e24b1592e060" (UID: "a567645b-5ae1-46a0-86b1-e24b1592e060"). InnerVolumeSpecName "kube-api-access-dv5wc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.480610 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a567645b-5ae1-46a0-86b1-e24b1592e060-scripts" (OuterVolumeSpecName: "scripts") pod "a567645b-5ae1-46a0-86b1-e24b1592e060" (UID: "a567645b-5ae1-46a0-86b1-e24b1592e060"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.487228 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a567645b-5ae1-46a0-86b1-e24b1592e060-config-data" (OuterVolumeSpecName: "config-data") pod "a567645b-5ae1-46a0-86b1-e24b1592e060" (UID: "a567645b-5ae1-46a0-86b1-e24b1592e060"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.535464 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a567645b-5ae1-46a0-86b1-e24b1592e060-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.535508 4578 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a567645b-5ae1-46a0-86b1-e24b1592e060-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.535524 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dv5wc\" (UniqueName: \"kubernetes.io/projected/a567645b-5ae1-46a0-86b1-e24b1592e060-kube-api-access-dv5wc\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.535535 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a567645b-5ae1-46a0-86b1-e24b1592e060-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.535545 4578 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a567645b-5ae1-46a0-86b1-e24b1592e060-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.591111 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.646939 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c684ff4d9-2h9cq" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.648670 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.744233 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-config\") pod \"b85d10a6-888d-4862-986f-5e44b8d3d229\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.744388 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-ovsdbserver-nb\") pod \"b85d10a6-888d-4862-986f-5e44b8d3d229\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.744467 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-dns-svc\") pod \"b85d10a6-888d-4862-986f-5e44b8d3d229\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.744497 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-dns-swift-storage-0\") pod \"b85d10a6-888d-4862-986f-5e44b8d3d229\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.744521 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnqwm\" (UniqueName: \"kubernetes.io/projected/b85d10a6-888d-4862-986f-5e44b8d3d229-kube-api-access-fnqwm\") pod \"b85d10a6-888d-4862-986f-5e44b8d3d229\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.744681 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-ovsdbserver-sb\") pod \"b85d10a6-888d-4862-986f-5e44b8d3d229\" (UID: \"b85d10a6-888d-4862-986f-5e44b8d3d229\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.779863 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b85d10a6-888d-4862-986f-5e44b8d3d229-kube-api-access-fnqwm" (OuterVolumeSpecName: "kube-api-access-fnqwm") pod "b85d10a6-888d-4862-986f-5e44b8d3d229" (UID: "b85d10a6-888d-4862-986f-5e44b8d3d229"). InnerVolumeSpecName "kube-api-access-fnqwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.821517 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b85d10a6-888d-4862-986f-5e44b8d3d229" (UID: "b85d10a6-888d-4862-986f-5e44b8d3d229"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.841877 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b85d10a6-888d-4862-986f-5e44b8d3d229" (UID: "b85d10a6-888d-4862-986f-5e44b8d3d229"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.845970 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-config\") pod \"101089c9-6f91-42fb-a5ab-d907c80ff069\" (UID: \"101089c9-6f91-42fb-a5ab-d907c80ff069\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.847250 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfkx6\" (UniqueName: \"kubernetes.io/projected/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-kube-api-access-gfkx6\") pod \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\" (UID: \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.847371 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-logs\") pod \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\" (UID: \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.847451 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-combined-ca-bundle\") pod \"101089c9-6f91-42fb-a5ab-d907c80ff069\" (UID: \"101089c9-6f91-42fb-a5ab-d907c80ff069\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.847581 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-scripts\") pod \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\" (UID: \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.847699 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-httpd-config\") pod \"101089c9-6f91-42fb-a5ab-d907c80ff069\" (UID: \"101089c9-6f91-42fb-a5ab-d907c80ff069\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.847775 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2n5t\" (UniqueName: \"kubernetes.io/projected/101089c9-6f91-42fb-a5ab-d907c80ff069-kube-api-access-s2n5t\") pod \"101089c9-6f91-42fb-a5ab-d907c80ff069\" (UID: \"101089c9-6f91-42fb-a5ab-d907c80ff069\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.847859 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-horizon-secret-key\") pod \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\" (UID: \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.847956 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-ovndb-tls-certs\") pod \"101089c9-6f91-42fb-a5ab-d907c80ff069\" (UID: \"101089c9-6f91-42fb-a5ab-d907c80ff069\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.848042 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-config-data\") pod \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\" (UID: \"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b\") " Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.848313 4578 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.848908 4578 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.848989 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnqwm\" (UniqueName: \"kubernetes.io/projected/b85d10a6-888d-4862-986f-5e44b8d3d229-kube-api-access-fnqwm\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.848654 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-logs" (OuterVolumeSpecName: "logs") pod "1b0f9e3d-a619-437f-9339-d61b6ffb8d4b" (UID: "1b0f9e3d-a619-437f-9339-d61b6ffb8d4b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.861786 4578 generic.go:334] "Generic (PLEG): container finished" podID="6228c849-5824-4a21-8318-754c65d79a2e" containerID="ed71ed5afef9b00e0ea00c8b27a6f35651ee63ca3bdc776cb10be4964a07b008" exitCode=2 Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.861868 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6228c849-5824-4a21-8318-754c65d79a2e","Type":"ContainerDied","Data":"ed71ed5afef9b00e0ea00c8b27a6f35651ee63ca3bdc776cb10be4964a07b008"} Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.867370 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c684ff4d9-2h9cq" event={"ID":"1b0f9e3d-a619-437f-9339-d61b6ffb8d4b","Type":"ContainerDied","Data":"14f3997b5a24d7cbcf43bc79436222ffecd03e2e5d33f7a0cd5061909e4ea151"} Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.867412 4578 scope.go:117] "RemoveContainer" containerID="80929f7ff9f2db6421b73861631c698d81ca2c15a2ea97522dc9d1eb510196b2" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.867525 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c684ff4d9-2h9cq" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.871680 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b85d10a6-888d-4862-986f-5e44b8d3d229" (UID: "b85d10a6-888d-4862-986f-5e44b8d3d229"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.877864 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "1b0f9e3d-a619-437f-9339-d61b6ffb8d4b" (UID: "1b0f9e3d-a619-437f-9339-d61b6ffb8d4b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.888001 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" event={"ID":"b85d10a6-888d-4862-986f-5e44b8d3d229","Type":"ContainerDied","Data":"5ef03001cc424673a570bab42eb85076c5d758abfe4c48e50a9378476d665592"} Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.888100 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-phtfx" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.889291 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-kube-api-access-gfkx6" (OuterVolumeSpecName: "kube-api-access-gfkx6") pod "1b0f9e3d-a619-437f-9339-d61b6ffb8d4b" (UID: "1b0f9e3d-a619-437f-9339-d61b6ffb8d4b"). InnerVolumeSpecName "kube-api-access-gfkx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.891774 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6d67444cbd-9qgkn" event={"ID":"101089c9-6f91-42fb-a5ab-d907c80ff069","Type":"ContainerDied","Data":"bbc5a55643316bbce08815fc7089ca7afc837fdc9726a9b22385bbe451a8e6fe"} Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.891918 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6d67444cbd-9qgkn" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.906709 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-config" (OuterVolumeSpecName: "config") pod "b85d10a6-888d-4862-986f-5e44b8d3d229" (UID: "b85d10a6-888d-4862-986f-5e44b8d3d229"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.910149 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "101089c9-6f91-42fb-a5ab-d907c80ff069" (UID: "101089c9-6f91-42fb-a5ab-d907c80ff069"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.910347 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-q6ppj" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.910437 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d5f486769-rxsds" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.910727 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d5f486769-rxsds" event={"ID":"a567645b-5ae1-46a0-86b1-e24b1592e060","Type":"ContainerDied","Data":"12363d6128e3c56437080d84701c50a13a9ab3b59cfcfd2c71de90be08a13a2b"} Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.924992 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/101089c9-6f91-42fb-a5ab-d907c80ff069-kube-api-access-s2n5t" (OuterVolumeSpecName: "kube-api-access-s2n5t") pod "101089c9-6f91-42fb-a5ab-d907c80ff069" (UID: "101089c9-6f91-42fb-a5ab-d907c80ff069"). InnerVolumeSpecName "kube-api-access-s2n5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.925493 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b85d10a6-888d-4862-986f-5e44b8d3d229" (UID: "b85d10a6-888d-4862-986f-5e44b8d3d229"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.951046 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.951075 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfkx6\" (UniqueName: \"kubernetes.io/projected/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-kube-api-access-gfkx6\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.951091 4578 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.951104 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.951115 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b85d10a6-888d-4862-986f-5e44b8d3d229-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.951126 4578 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.951136 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2n5t\" (UniqueName: \"kubernetes.io/projected/101089c9-6f91-42fb-a5ab-d907c80ff069-kube-api-access-s2n5t\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.951147 4578 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.978613 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-scripts" (OuterVolumeSpecName: "scripts") pod "1b0f9e3d-a619-437f-9339-d61b6ffb8d4b" (UID: "1b0f9e3d-a619-437f-9339-d61b6ffb8d4b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.981926 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-config" (OuterVolumeSpecName: "config") pod "101089c9-6f91-42fb-a5ab-d907c80ff069" (UID: "101089c9-6f91-42fb-a5ab-d907c80ff069"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:21 crc kubenswrapper[4578]: I1003 13:13:21.984202 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-config-data" (OuterVolumeSpecName: "config-data") pod "1b0f9e3d-a619-437f-9339-d61b6ffb8d4b" (UID: "1b0f9e3d-a619-437f-9339-d61b6ffb8d4b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.054799 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.054830 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.054843 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.055888 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "101089c9-6f91-42fb-a5ab-d907c80ff069" (UID: "101089c9-6f91-42fb-a5ab-d907c80ff069"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.063292 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d5f486769-rxsds"] Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.073817 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "101089c9-6f91-42fb-a5ab-d907c80ff069" (UID: "101089c9-6f91-42fb-a5ab-d907c80ff069"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.075833 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-d5f486769-rxsds"] Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.162682 4578 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.162722 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/101089c9-6f91-42fb-a5ab-d907c80ff069-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.190662 4578 scope.go:117] "RemoveContainer" containerID="5402c4c9ddc18e14fe296ec7864a7e4a97e4da6535530acc67d68775074d1046" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.226209 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c684ff4d9-2h9cq"] Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.251872 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7c684ff4d9-2h9cq"] Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.294032 4578 scope.go:117] "RemoveContainer" containerID="460acd0e4b7cdbdda5a53c34a461e8bf587fbb22d1ee478ae0ec8f7d216a1abc" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.314280 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-phtfx"] Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.334126 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-phtfx"] Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.345797 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:13:22 crc kubenswrapper[4578]: E1003 13:13:22.346226 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b85d10a6-888d-4862-986f-5e44b8d3d229" containerName="init" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.346242 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="b85d10a6-888d-4862-986f-5e44b8d3d229" containerName="init" Oct 03 13:13:22 crc kubenswrapper[4578]: E1003 13:13:22.346259 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b0f9e3d-a619-437f-9339-d61b6ffb8d4b" containerName="horizon-log" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.346265 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b0f9e3d-a619-437f-9339-d61b6ffb8d4b" containerName="horizon-log" Oct 03 13:13:22 crc kubenswrapper[4578]: E1003 13:13:22.346288 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b0f9e3d-a619-437f-9339-d61b6ffb8d4b" containerName="horizon" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.346293 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b0f9e3d-a619-437f-9339-d61b6ffb8d4b" containerName="horizon" Oct 03 13:13:22 crc kubenswrapper[4578]: E1003 13:13:22.346303 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b310034-f02f-48ac-a47a-4f41f15379a5" containerName="cinder-db-sync" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.346308 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b310034-f02f-48ac-a47a-4f41f15379a5" containerName="cinder-db-sync" Oct 03 13:13:22 crc kubenswrapper[4578]: E1003 13:13:22.346316 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b85d10a6-888d-4862-986f-5e44b8d3d229" containerName="dnsmasq-dns" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.346321 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="b85d10a6-888d-4862-986f-5e44b8d3d229" containerName="dnsmasq-dns" Oct 03 13:13:22 crc kubenswrapper[4578]: E1003 13:13:22.346330 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="101089c9-6f91-42fb-a5ab-d907c80ff069" containerName="neutron-httpd" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.346338 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="101089c9-6f91-42fb-a5ab-d907c80ff069" containerName="neutron-httpd" Oct 03 13:13:22 crc kubenswrapper[4578]: E1003 13:13:22.346364 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a567645b-5ae1-46a0-86b1-e24b1592e060" containerName="horizon" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.346370 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a567645b-5ae1-46a0-86b1-e24b1592e060" containerName="horizon" Oct 03 13:13:22 crc kubenswrapper[4578]: E1003 13:13:22.346385 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a567645b-5ae1-46a0-86b1-e24b1592e060" containerName="horizon-log" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.346391 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a567645b-5ae1-46a0-86b1-e24b1592e060" containerName="horizon-log" Oct 03 13:13:22 crc kubenswrapper[4578]: E1003 13:13:22.346399 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="101089c9-6f91-42fb-a5ab-d907c80ff069" containerName="neutron-api" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.346405 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="101089c9-6f91-42fb-a5ab-d907c80ff069" containerName="neutron-api" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.346575 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="a567645b-5ae1-46a0-86b1-e24b1592e060" containerName="horizon-log" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.346585 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="a567645b-5ae1-46a0-86b1-e24b1592e060" containerName="horizon" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.346594 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b310034-f02f-48ac-a47a-4f41f15379a5" containerName="cinder-db-sync" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.346607 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b0f9e3d-a619-437f-9339-d61b6ffb8d4b" containerName="horizon" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.346617 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="101089c9-6f91-42fb-a5ab-d907c80ff069" containerName="neutron-httpd" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.346645 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="101089c9-6f91-42fb-a5ab-d907c80ff069" containerName="neutron-api" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.346651 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="b85d10a6-888d-4862-986f-5e44b8d3d229" containerName="dnsmasq-dns" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.346658 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b0f9e3d-a619-437f-9339-d61b6ffb8d4b" containerName="horizon-log" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.347566 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.378869 4578 scope.go:117] "RemoveContainer" containerID="90e2e3f7ccedd25cf461c8c8da59d658d9b7cc0793e9947c3ceafc7418d8ad90" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.379457 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.379714 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.380419 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-p5nqp" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.380587 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.409300 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6d67444cbd-9qgkn"] Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.444105 4578 scope.go:117] "RemoveContainer" containerID="6617906f8965d5b3d283bc14710803835bf60a5cb3e056a9c10f0e336b4cdff7" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.450395 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6d67444cbd-9qgkn"] Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.480842 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.486003 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-config-data\") pod \"cinder-scheduler-0\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.486056 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-scripts\") pod \"cinder-scheduler-0\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.486107 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.486170 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.486259 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aad29d7f-9721-491d-b3b6-5064b953ac45-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.486289 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5tvj\" (UniqueName: \"kubernetes.io/projected/aad29d7f-9721-491d-b3b6-5064b953ac45-kube-api-access-k5tvj\") pod \"cinder-scheduler-0\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.489499 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-45rx4"] Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.491771 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.501505 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-45rx4"] Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.513554 4578 scope.go:117] "RemoveContainer" containerID="e73cf2c70f9bc072fedc51586fa3a7d7b4c59719b48f280707a595fd70f5d102" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.579844 4578 scope.go:117] "RemoveContainer" containerID="4b3f4f06273a71501b5530c10beab91d989254af38cef506f55056b9e26f0b06" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.592686 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.593556 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.593616 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-45rx4\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.593665 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aad29d7f-9721-491d-b3b6-5064b953ac45-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.593688 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5tvj\" (UniqueName: \"kubernetes.io/projected/aad29d7f-9721-491d-b3b6-5064b953ac45-kube-api-access-k5tvj\") pod \"cinder-scheduler-0\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.593724 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-dns-svc\") pod \"dnsmasq-dns-5784cf869f-45rx4\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.593769 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwgfv\" (UniqueName: \"kubernetes.io/projected/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-kube-api-access-pwgfv\") pod \"dnsmasq-dns-5784cf869f-45rx4\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.593787 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-config\") pod \"dnsmasq-dns-5784cf869f-45rx4\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.593825 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-config-data\") pod \"cinder-scheduler-0\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.593845 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-scripts\") pod \"cinder-scheduler-0\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.593863 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-45rx4\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.593881 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-45rx4\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.593905 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.594325 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.593781 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aad29d7f-9721-491d-b3b6-5064b953ac45-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.600020 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.602475 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-config-data\") pod \"cinder-scheduler-0\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.603582 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.605694 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.606139 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.607171 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-scripts\") pod \"cinder-scheduler-0\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.631435 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5tvj\" (UniqueName: \"kubernetes.io/projected/aad29d7f-9721-491d-b3b6-5064b953ac45-kube-api-access-k5tvj\") pod \"cinder-scheduler-0\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.695519 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-45rx4\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.695578 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.695660 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-dns-svc\") pod \"dnsmasq-dns-5784cf869f-45rx4\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.695732 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwgfv\" (UniqueName: \"kubernetes.io/projected/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-kube-api-access-pwgfv\") pod \"dnsmasq-dns-5784cf869f-45rx4\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.695751 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-scripts\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.695769 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-config\") pod \"dnsmasq-dns-5784cf869f-45rx4\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.695804 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-config-data-custom\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.695824 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-45rx4\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.695841 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-config-data\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.695854 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-logs\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.695889 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-45rx4\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.695909 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g6xk\" (UniqueName: \"kubernetes.io/projected/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-kube-api-access-2g6xk\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.695928 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.696880 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-dns-svc\") pod \"dnsmasq-dns-5784cf869f-45rx4\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.697525 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-45rx4\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.697801 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-config\") pod \"dnsmasq-dns-5784cf869f-45rx4\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.699118 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-45rx4\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.700590 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-45rx4\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.720154 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwgfv\" (UniqueName: \"kubernetes.io/projected/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-kube-api-access-pwgfv\") pod \"dnsmasq-dns-5784cf869f-45rx4\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.746971 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.797993 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-scripts\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.798040 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-config-data-custom\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.798065 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-config-data\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.798083 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-logs\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.798106 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g6xk\" (UniqueName: \"kubernetes.io/projected/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-kube-api-access-2g6xk\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.798126 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.798204 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.798686 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.798962 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-logs\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.804084 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-config-data-custom\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.805834 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-scripts\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.806151 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.826760 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.833852 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-config-data\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.855440 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g6xk\" (UniqueName: \"kubernetes.io/projected/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-kube-api-access-2g6xk\") pod \"cinder-api-0\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " pod="openstack/cinder-api-0" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.894464 4578 scope.go:117] "RemoveContainer" containerID="00d4bed9ee5fe90eebe7ce2ec9be50c6902436a20863280d15aafd3f0158f87d" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.931367 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="101089c9-6f91-42fb-a5ab-d907c80ff069" path="/var/lib/kubelet/pods/101089c9-6f91-42fb-a5ab-d907c80ff069/volumes" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.932793 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b0f9e3d-a619-437f-9339-d61b6ffb8d4b" path="/var/lib/kubelet/pods/1b0f9e3d-a619-437f-9339-d61b6ffb8d4b/volumes" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.933768 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a567645b-5ae1-46a0-86b1-e24b1592e060" path="/var/lib/kubelet/pods/a567645b-5ae1-46a0-86b1-e24b1592e060/volumes" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.935339 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b85d10a6-888d-4862-986f-5e44b8d3d229" path="/var/lib/kubelet/pods/b85d10a6-888d-4862-986f-5e44b8d3d229/volumes" Oct 03 13:13:22 crc kubenswrapper[4578]: I1003 13:13:22.948962 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 13:13:23 crc kubenswrapper[4578]: I1003 13:13:23.192438 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-85d5675966-rb6nf" podUID="b40b382f-7fb7-4f67-9237-e5148abf7610" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:53204->10.217.0.158:9311: read: connection reset by peer" Oct 03 13:13:23 crc kubenswrapper[4578]: I1003 13:13:23.192498 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-85d5675966-rb6nf" podUID="b40b382f-7fb7-4f67-9237-e5148abf7610" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:53210->10.217.0.158:9311: read: connection reset by peer" Oct 03 13:13:23 crc kubenswrapper[4578]: I1003 13:13:23.193131 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:13:23 crc kubenswrapper[4578]: I1003 13:13:23.622371 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:13:23 crc kubenswrapper[4578]: I1003 13:13:23.657964 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:13:23 crc kubenswrapper[4578]: W1003 13:13:23.669026 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4aba55d1_ad9e_4803_b2ed_ff0fdd6f4f0e.slice/crio-9683bc625846b05c479603f46c56653c3ffd6077680c9d30c0b38c1cf9e1046d WatchSource:0}: Error finding container 9683bc625846b05c479603f46c56653c3ffd6077680c9d30c0b38c1cf9e1046d: Status 404 returned error can't find the container with id 9683bc625846b05c479603f46c56653c3ffd6077680c9d30c0b38c1cf9e1046d Oct 03 13:13:23 crc kubenswrapper[4578]: I1003 13:13:23.743309 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-45rx4"] Oct 03 13:13:23 crc kubenswrapper[4578]: I1003 13:13:23.903918 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:13:23 crc kubenswrapper[4578]: I1003 13:13:23.984203 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"aad29d7f-9721-491d-b3b6-5064b953ac45","Type":"ContainerStarted","Data":"87e8c3706d4377f934b54e5c7c1d08e05d3151a52cee54923322e13e48e518e1"} Oct 03 13:13:23 crc kubenswrapper[4578]: I1003 13:13:23.985821 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-45rx4" event={"ID":"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f","Type":"ContainerStarted","Data":"521b7ef2049075390765ad2edf95200d4dc5c8e1ab2374715bbf5d59e51114ab"} Oct 03 13:13:23 crc kubenswrapper[4578]: I1003 13:13:23.997816 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e","Type":"ContainerStarted","Data":"9683bc625846b05c479603f46c56653c3ffd6077680c9d30c0b38c1cf9e1046d"} Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.018452 4578 generic.go:334] "Generic (PLEG): container finished" podID="b40b382f-7fb7-4f67-9237-e5148abf7610" containerID="9bfdb147e2aee44432322ce8075c6137266854d584755636f7e42e56c26ab192" exitCode=0 Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.018508 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85d5675966-rb6nf" event={"ID":"b40b382f-7fb7-4f67-9237-e5148abf7610","Type":"ContainerDied","Data":"9bfdb147e2aee44432322ce8075c6137266854d584755636f7e42e56c26ab192"} Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.018546 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-85d5675966-rb6nf" event={"ID":"b40b382f-7fb7-4f67-9237-e5148abf7610","Type":"ContainerDied","Data":"e37342f43e0cb2ae68aa5a669f4a4fe93deff06fcbc18c195742a65adb06a46b"} Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.018564 4578 scope.go:117] "RemoveContainer" containerID="9bfdb147e2aee44432322ce8075c6137266854d584755636f7e42e56c26ab192" Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.018696 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-85d5675966-rb6nf" Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.023846 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b40b382f-7fb7-4f67-9237-e5148abf7610-config-data\") pod \"b40b382f-7fb7-4f67-9237-e5148abf7610\" (UID: \"b40b382f-7fb7-4f67-9237-e5148abf7610\") " Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.023962 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b40b382f-7fb7-4f67-9237-e5148abf7610-config-data-custom\") pod \"b40b382f-7fb7-4f67-9237-e5148abf7610\" (UID: \"b40b382f-7fb7-4f67-9237-e5148abf7610\") " Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.023996 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b40b382f-7fb7-4f67-9237-e5148abf7610-combined-ca-bundle\") pod \"b40b382f-7fb7-4f67-9237-e5148abf7610\" (UID: \"b40b382f-7fb7-4f67-9237-e5148abf7610\") " Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.024062 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b40b382f-7fb7-4f67-9237-e5148abf7610-logs\") pod \"b40b382f-7fb7-4f67-9237-e5148abf7610\" (UID: \"b40b382f-7fb7-4f67-9237-e5148abf7610\") " Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.024208 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffqmv\" (UniqueName: \"kubernetes.io/projected/b40b382f-7fb7-4f67-9237-e5148abf7610-kube-api-access-ffqmv\") pod \"b40b382f-7fb7-4f67-9237-e5148abf7610\" (UID: \"b40b382f-7fb7-4f67-9237-e5148abf7610\") " Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.024742 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b40b382f-7fb7-4f67-9237-e5148abf7610-logs" (OuterVolumeSpecName: "logs") pod "b40b382f-7fb7-4f67-9237-e5148abf7610" (UID: "b40b382f-7fb7-4f67-9237-e5148abf7610"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.024908 4578 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b40b382f-7fb7-4f67-9237-e5148abf7610-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.065012 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b40b382f-7fb7-4f67-9237-e5148abf7610-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b40b382f-7fb7-4f67-9237-e5148abf7610" (UID: "b40b382f-7fb7-4f67-9237-e5148abf7610"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.084753 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b40b382f-7fb7-4f67-9237-e5148abf7610-kube-api-access-ffqmv" (OuterVolumeSpecName: "kube-api-access-ffqmv") pod "b40b382f-7fb7-4f67-9237-e5148abf7610" (UID: "b40b382f-7fb7-4f67-9237-e5148abf7610"). InnerVolumeSpecName "kube-api-access-ffqmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.130127 4578 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b40b382f-7fb7-4f67-9237-e5148abf7610-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.130152 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffqmv\" (UniqueName: \"kubernetes.io/projected/b40b382f-7fb7-4f67-9237-e5148abf7610-kube-api-access-ffqmv\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.199970 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b40b382f-7fb7-4f67-9237-e5148abf7610-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b40b382f-7fb7-4f67-9237-e5148abf7610" (UID: "b40b382f-7fb7-4f67-9237-e5148abf7610"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.208362 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.235971 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b40b382f-7fb7-4f67-9237-e5148abf7610-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.257990 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b40b382f-7fb7-4f67-9237-e5148abf7610-config-data" (OuterVolumeSpecName: "config-data") pod "b40b382f-7fb7-4f67-9237-e5148abf7610" (UID: "b40b382f-7fb7-4f67-9237-e5148abf7610"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.271557 4578 scope.go:117] "RemoveContainer" containerID="61dc6fff53873a9950a28c8817606b415df204b784b5fccbc3d21327dc5f282d" Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.327795 4578 scope.go:117] "RemoveContainer" containerID="9bfdb147e2aee44432322ce8075c6137266854d584755636f7e42e56c26ab192" Oct 03 13:13:24 crc kubenswrapper[4578]: E1003 13:13:24.328438 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bfdb147e2aee44432322ce8075c6137266854d584755636f7e42e56c26ab192\": container with ID starting with 9bfdb147e2aee44432322ce8075c6137266854d584755636f7e42e56c26ab192 not found: ID does not exist" containerID="9bfdb147e2aee44432322ce8075c6137266854d584755636f7e42e56c26ab192" Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.329673 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bfdb147e2aee44432322ce8075c6137266854d584755636f7e42e56c26ab192"} err="failed to get container status \"9bfdb147e2aee44432322ce8075c6137266854d584755636f7e42e56c26ab192\": rpc error: code = NotFound desc = could not find container \"9bfdb147e2aee44432322ce8075c6137266854d584755636f7e42e56c26ab192\": container with ID starting with 9bfdb147e2aee44432322ce8075c6137266854d584755636f7e42e56c26ab192 not found: ID does not exist" Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.329886 4578 scope.go:117] "RemoveContainer" containerID="61dc6fff53873a9950a28c8817606b415df204b784b5fccbc3d21327dc5f282d" Oct 03 13:13:24 crc kubenswrapper[4578]: E1003 13:13:24.331779 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61dc6fff53873a9950a28c8817606b415df204b784b5fccbc3d21327dc5f282d\": container with ID starting with 61dc6fff53873a9950a28c8817606b415df204b784b5fccbc3d21327dc5f282d not found: ID does not exist" containerID="61dc6fff53873a9950a28c8817606b415df204b784b5fccbc3d21327dc5f282d" Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.331912 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61dc6fff53873a9950a28c8817606b415df204b784b5fccbc3d21327dc5f282d"} err="failed to get container status \"61dc6fff53873a9950a28c8817606b415df204b784b5fccbc3d21327dc5f282d\": rpc error: code = NotFound desc = could not find container \"61dc6fff53873a9950a28c8817606b415df204b784b5fccbc3d21327dc5f282d\": container with ID starting with 61dc6fff53873a9950a28c8817606b415df204b784b5fccbc3d21327dc5f282d not found: ID does not exist" Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.332176 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-66cbcb46d8-9jksg" Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.339373 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b40b382f-7fb7-4f67-9237-e5148abf7610-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.440705 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-85d5675966-rb6nf"] Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.503862 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-85d5675966-rb6nf"] Oct 03 13:13:24 crc kubenswrapper[4578]: I1003 13:13:24.933613 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b40b382f-7fb7-4f67-9237-e5148abf7610" path="/var/lib/kubelet/pods/b40b382f-7fb7-4f67-9237-e5148abf7610/volumes" Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.056470 4578 generic.go:334] "Generic (PLEG): container finished" podID="7e0fa932-47ae-4ed6-94b3-0b2829dbab5f" containerID="1556a725891caf1302c447126133d7a876f00aff2ac0a26aca363a8f02981637" exitCode=0 Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.056591 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-45rx4" event={"ID":"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f","Type":"ContainerDied","Data":"1556a725891caf1302c447126133d7a876f00aff2ac0a26aca363a8f02981637"} Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.068994 4578 generic.go:334] "Generic (PLEG): container finished" podID="6228c849-5824-4a21-8318-754c65d79a2e" containerID="b45d39e61e8fc0942f781f361e7f58c723585ade153a008212d11232868eaf4d" exitCode=0 Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.069077 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6228c849-5824-4a21-8318-754c65d79a2e","Type":"ContainerDied","Data":"b45d39e61e8fc0942f781f361e7f58c723585ade153a008212d11232868eaf4d"} Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.195223 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.353967 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-config-data\") pod \"6228c849-5824-4a21-8318-754c65d79a2e\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.354007 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbm4c\" (UniqueName: \"kubernetes.io/projected/6228c849-5824-4a21-8318-754c65d79a2e-kube-api-access-tbm4c\") pod \"6228c849-5824-4a21-8318-754c65d79a2e\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.354067 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-scripts\") pod \"6228c849-5824-4a21-8318-754c65d79a2e\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.354535 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-combined-ca-bundle\") pod \"6228c849-5824-4a21-8318-754c65d79a2e\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.354651 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-sg-core-conf-yaml\") pod \"6228c849-5824-4a21-8318-754c65d79a2e\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.354740 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6228c849-5824-4a21-8318-754c65d79a2e-log-httpd\") pod \"6228c849-5824-4a21-8318-754c65d79a2e\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.354772 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6228c849-5824-4a21-8318-754c65d79a2e-run-httpd\") pod \"6228c849-5824-4a21-8318-754c65d79a2e\" (UID: \"6228c849-5824-4a21-8318-754c65d79a2e\") " Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.355208 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6228c849-5824-4a21-8318-754c65d79a2e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6228c849-5824-4a21-8318-754c65d79a2e" (UID: "6228c849-5824-4a21-8318-754c65d79a2e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.355532 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6228c849-5824-4a21-8318-754c65d79a2e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6228c849-5824-4a21-8318-754c65d79a2e" (UID: "6228c849-5824-4a21-8318-754c65d79a2e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.369858 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-scripts" (OuterVolumeSpecName: "scripts") pod "6228c849-5824-4a21-8318-754c65d79a2e" (UID: "6228c849-5824-4a21-8318-754c65d79a2e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.369986 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6228c849-5824-4a21-8318-754c65d79a2e-kube-api-access-tbm4c" (OuterVolumeSpecName: "kube-api-access-tbm4c") pod "6228c849-5824-4a21-8318-754c65d79a2e" (UID: "6228c849-5824-4a21-8318-754c65d79a2e"). InnerVolumeSpecName "kube-api-access-tbm4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.389797 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-config-data" (OuterVolumeSpecName: "config-data") pod "6228c849-5824-4a21-8318-754c65d79a2e" (UID: "6228c849-5824-4a21-8318-754c65d79a2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.395004 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6228c849-5824-4a21-8318-754c65d79a2e" (UID: "6228c849-5824-4a21-8318-754c65d79a2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.398768 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6228c849-5824-4a21-8318-754c65d79a2e" (UID: "6228c849-5824-4a21-8318-754c65d79a2e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.459301 4578 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.459332 4578 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6228c849-5824-4a21-8318-754c65d79a2e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.459341 4578 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6228c849-5824-4a21-8318-754c65d79a2e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.459348 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.459360 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbm4c\" (UniqueName: \"kubernetes.io/projected/6228c849-5824-4a21-8318-754c65d79a2e-kube-api-access-tbm4c\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.459368 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.459376 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6228c849-5824-4a21-8318-754c65d79a2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:25 crc kubenswrapper[4578]: I1003 13:13:25.661471 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.106021 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-45rx4" event={"ID":"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f","Type":"ContainerStarted","Data":"40ae7f33bb143edc84844d6355501b2b08b9e2e05d72567ae4a58998411b19b5"} Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.106527 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.124199 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e","Type":"ContainerStarted","Data":"fa312bfcc4c7878cb4aa4e823d7f08cb05d5351b8b33901207596e9c69d6fbb2"} Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.130479 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6228c849-5824-4a21-8318-754c65d79a2e","Type":"ContainerDied","Data":"848c98d057abb33e3085916996b02be877d7016c169a653c7e1a52f0bf10b9f2"} Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.130552 4578 scope.go:117] "RemoveContainer" containerID="ed71ed5afef9b00e0ea00c8b27a6f35651ee63ca3bdc776cb10be4964a07b008" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.130770 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.142308 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"aad29d7f-9721-491d-b3b6-5064b953ac45","Type":"ContainerStarted","Data":"34bfe862b66fd73388fa9e7185e5cbc50746bca0f3cc35a05193204a1a27aa5a"} Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.145274 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-45rx4" podStartSLOduration=4.145256488 podStartE2EDuration="4.145256488s" podCreationTimestamp="2025-10-03 13:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:13:26.132317905 +0000 UTC m=+1341.930790109" watchObservedRunningTime="2025-10-03 13:13:26.145256488 +0000 UTC m=+1341.943728672" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.235272 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.235387 4578 scope.go:117] "RemoveContainer" containerID="b45d39e61e8fc0942f781f361e7f58c723585ade153a008212d11232868eaf4d" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.263570 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.300513 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:13:26 crc kubenswrapper[4578]: E1003 13:13:26.300966 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40b382f-7fb7-4f67-9237-e5148abf7610" containerName="barbican-api-log" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.300984 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40b382f-7fb7-4f67-9237-e5148abf7610" containerName="barbican-api-log" Oct 03 13:13:26 crc kubenswrapper[4578]: E1003 13:13:26.300995 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6228c849-5824-4a21-8318-754c65d79a2e" containerName="ceilometer-notification-agent" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.301002 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="6228c849-5824-4a21-8318-754c65d79a2e" containerName="ceilometer-notification-agent" Oct 03 13:13:26 crc kubenswrapper[4578]: E1003 13:13:26.301024 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b40b382f-7fb7-4f67-9237-e5148abf7610" containerName="barbican-api" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.301032 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="b40b382f-7fb7-4f67-9237-e5148abf7610" containerName="barbican-api" Oct 03 13:13:26 crc kubenswrapper[4578]: E1003 13:13:26.301053 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6228c849-5824-4a21-8318-754c65d79a2e" containerName="sg-core" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.301059 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="6228c849-5824-4a21-8318-754c65d79a2e" containerName="sg-core" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.301272 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40b382f-7fb7-4f67-9237-e5148abf7610" containerName="barbican-api" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.301288 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="6228c849-5824-4a21-8318-754c65d79a2e" containerName="sg-core" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.301311 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="b40b382f-7fb7-4f67-9237-e5148abf7610" containerName="barbican-api-log" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.301320 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="6228c849-5824-4a21-8318-754c65d79a2e" containerName="ceilometer-notification-agent" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.308710 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.313166 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.313352 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.318127 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.480450 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbbb1889-c411-41c1-81f2-ee4ab509acc9-run-httpd\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.480535 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.480585 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-scripts\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.480609 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbbb1889-c411-41c1-81f2-ee4ab509acc9-log-httpd\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.480656 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdccd\" (UniqueName: \"kubernetes.io/projected/dbbb1889-c411-41c1-81f2-ee4ab509acc9-kube-api-access-rdccd\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.480672 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-config-data\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.480881 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.581956 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbbb1889-c411-41c1-81f2-ee4ab509acc9-run-httpd\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.582020 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.582413 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbbb1889-c411-41c1-81f2-ee4ab509acc9-run-httpd\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.582963 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-scripts\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.583000 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbbb1889-c411-41c1-81f2-ee4ab509acc9-log-httpd\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.583030 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-config-data\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.583046 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdccd\" (UniqueName: \"kubernetes.io/projected/dbbb1889-c411-41c1-81f2-ee4ab509acc9-kube-api-access-rdccd\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.583098 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.583384 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbbb1889-c411-41c1-81f2-ee4ab509acc9-log-httpd\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.588129 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-scripts\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.589188 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.598765 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-config-data\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.603139 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.607278 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdccd\" (UniqueName: \"kubernetes.io/projected/dbbb1889-c411-41c1-81f2-ee4ab509acc9-kube-api-access-rdccd\") pod \"ceilometer-0\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.657336 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:13:26 crc kubenswrapper[4578]: I1003 13:13:26.991004 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6228c849-5824-4a21-8318-754c65d79a2e" path="/var/lib/kubelet/pods/6228c849-5824-4a21-8318-754c65d79a2e/volumes" Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.173937 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbbb1889-c411-41c1-81f2-ee4ab509acc9","Type":"ContainerStarted","Data":"38bb1d9921b1f9d83a95bcfbab52467ba6b8287e14cdd572fc44e923def3821d"} Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.179532 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.183267 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e","Type":"ContainerStarted","Data":"c0d8fa6fd8eab8093803f508adf6df30d5a1fba8276f851dd3693b9058331b05"} Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.183366 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e" containerName="cinder-api-log" containerID="cri-o://fa312bfcc4c7878cb4aa4e823d7f08cb05d5351b8b33901207596e9c69d6fbb2" gracePeriod=30 Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.183455 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.183483 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e" containerName="cinder-api" containerID="cri-o://c0d8fa6fd8eab8093803f508adf6df30d5a1fba8276f851dd3693b9058331b05" gracePeriod=30 Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.203356 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"aad29d7f-9721-491d-b3b6-5064b953ac45","Type":"ContainerStarted","Data":"bd3d34b2b6f782764b77d72f37586480d926b7aa8b52a17b6f5abe0621841a8d"} Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.214797 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.214780516 podStartE2EDuration="5.214780516s" podCreationTimestamp="2025-10-03 13:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:13:27.20301448 +0000 UTC m=+1343.001486664" watchObservedRunningTime="2025-10-03 13:13:27.214780516 +0000 UTC m=+1343.013252700" Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.235132 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.97815514 podStartE2EDuration="5.23511479s" podCreationTimestamp="2025-10-03 13:13:22 +0000 UTC" firstStartedPulling="2025-10-03 13:13:23.66125139 +0000 UTC m=+1339.459723574" lastFinishedPulling="2025-10-03 13:13:24.91821104 +0000 UTC m=+1340.716683224" observedRunningTime="2025-10-03 13:13:27.23444472 +0000 UTC m=+1343.032916904" watchObservedRunningTime="2025-10-03 13:13:27.23511479 +0000 UTC m=+1343.033586974" Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.747743 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.862674 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.916335 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-config-data-custom\") pod \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.916407 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-config-data\") pod \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.916448 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2g6xk\" (UniqueName: \"kubernetes.io/projected/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-kube-api-access-2g6xk\") pod \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.916494 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-combined-ca-bundle\") pod \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.916511 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-etc-machine-id\") pod \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.916608 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-scripts\") pod \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.916658 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-logs\") pod \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\" (UID: \"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e\") " Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.916735 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e" (UID: "4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.917067 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-logs" (OuterVolumeSpecName: "logs") pod "4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e" (UID: "4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.917884 4578 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.917911 4578 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.922286 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e" (UID: "4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.932945 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-kube-api-access-2g6xk" (OuterVolumeSpecName: "kube-api-access-2g6xk") pod "4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e" (UID: "4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e"). InnerVolumeSpecName "kube-api-access-2g6xk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.937095 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-scripts" (OuterVolumeSpecName: "scripts") pod "4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e" (UID: "4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:27 crc kubenswrapper[4578]: I1003 13:13:27.954013 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e" (UID: "4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.019481 4578 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.019703 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2g6xk\" (UniqueName: \"kubernetes.io/projected/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-kube-api-access-2g6xk\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.019809 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.019878 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.062878 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-config-data" (OuterVolumeSpecName: "config-data") pod "4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e" (UID: "4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.121668 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.215982 4578 generic.go:334] "Generic (PLEG): container finished" podID="4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e" containerID="c0d8fa6fd8eab8093803f508adf6df30d5a1fba8276f851dd3693b9058331b05" exitCode=0 Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.216285 4578 generic.go:334] "Generic (PLEG): container finished" podID="4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e" containerID="fa312bfcc4c7878cb4aa4e823d7f08cb05d5351b8b33901207596e9c69d6fbb2" exitCode=143 Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.216104 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.216127 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e","Type":"ContainerDied","Data":"c0d8fa6fd8eab8093803f508adf6df30d5a1fba8276f851dd3693b9058331b05"} Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.216439 4578 scope.go:117] "RemoveContainer" containerID="c0d8fa6fd8eab8093803f508adf6df30d5a1fba8276f851dd3693b9058331b05" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.216416 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e","Type":"ContainerDied","Data":"fa312bfcc4c7878cb4aa4e823d7f08cb05d5351b8b33901207596e9c69d6fbb2"} Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.216867 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e","Type":"ContainerDied","Data":"9683bc625846b05c479603f46c56653c3ffd6077680c9d30c0b38c1cf9e1046d"} Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.255123 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.268647 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.270279 4578 scope.go:117] "RemoveContainer" containerID="fa312bfcc4c7878cb4aa4e823d7f08cb05d5351b8b33901207596e9c69d6fbb2" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.287532 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:13:28 crc kubenswrapper[4578]: E1003 13:13:28.287887 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e" containerName="cinder-api-log" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.287901 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e" containerName="cinder-api-log" Oct 03 13:13:28 crc kubenswrapper[4578]: E1003 13:13:28.287920 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e" containerName="cinder-api" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.287926 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e" containerName="cinder-api" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.288102 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e" containerName="cinder-api-log" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.288116 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e" containerName="cinder-api" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.289002 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.294271 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.294397 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.300965 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.307891 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.316550 4578 scope.go:117] "RemoveContainer" containerID="c0d8fa6fd8eab8093803f508adf6df30d5a1fba8276f851dd3693b9058331b05" Oct 03 13:13:28 crc kubenswrapper[4578]: E1003 13:13:28.318061 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0d8fa6fd8eab8093803f508adf6df30d5a1fba8276f851dd3693b9058331b05\": container with ID starting with c0d8fa6fd8eab8093803f508adf6df30d5a1fba8276f851dd3693b9058331b05 not found: ID does not exist" containerID="c0d8fa6fd8eab8093803f508adf6df30d5a1fba8276f851dd3693b9058331b05" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.318101 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0d8fa6fd8eab8093803f508adf6df30d5a1fba8276f851dd3693b9058331b05"} err="failed to get container status \"c0d8fa6fd8eab8093803f508adf6df30d5a1fba8276f851dd3693b9058331b05\": rpc error: code = NotFound desc = could not find container \"c0d8fa6fd8eab8093803f508adf6df30d5a1fba8276f851dd3693b9058331b05\": container with ID starting with c0d8fa6fd8eab8093803f508adf6df30d5a1fba8276f851dd3693b9058331b05 not found: ID does not exist" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.318125 4578 scope.go:117] "RemoveContainer" containerID="fa312bfcc4c7878cb4aa4e823d7f08cb05d5351b8b33901207596e9c69d6fbb2" Oct 03 13:13:28 crc kubenswrapper[4578]: E1003 13:13:28.318454 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa312bfcc4c7878cb4aa4e823d7f08cb05d5351b8b33901207596e9c69d6fbb2\": container with ID starting with fa312bfcc4c7878cb4aa4e823d7f08cb05d5351b8b33901207596e9c69d6fbb2 not found: ID does not exist" containerID="fa312bfcc4c7878cb4aa4e823d7f08cb05d5351b8b33901207596e9c69d6fbb2" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.318479 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa312bfcc4c7878cb4aa4e823d7f08cb05d5351b8b33901207596e9c69d6fbb2"} err="failed to get container status \"fa312bfcc4c7878cb4aa4e823d7f08cb05d5351b8b33901207596e9c69d6fbb2\": rpc error: code = NotFound desc = could not find container \"fa312bfcc4c7878cb4aa4e823d7f08cb05d5351b8b33901207596e9c69d6fbb2\": container with ID starting with fa312bfcc4c7878cb4aa4e823d7f08cb05d5351b8b33901207596e9c69d6fbb2 not found: ID does not exist" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.318516 4578 scope.go:117] "RemoveContainer" containerID="c0d8fa6fd8eab8093803f508adf6df30d5a1fba8276f851dd3693b9058331b05" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.321969 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0d8fa6fd8eab8093803f508adf6df30d5a1fba8276f851dd3693b9058331b05"} err="failed to get container status \"c0d8fa6fd8eab8093803f508adf6df30d5a1fba8276f851dd3693b9058331b05\": rpc error: code = NotFound desc = could not find container \"c0d8fa6fd8eab8093803f508adf6df30d5a1fba8276f851dd3693b9058331b05\": container with ID starting with c0d8fa6fd8eab8093803f508adf6df30d5a1fba8276f851dd3693b9058331b05 not found: ID does not exist" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.321992 4578 scope.go:117] "RemoveContainer" containerID="fa312bfcc4c7878cb4aa4e823d7f08cb05d5351b8b33901207596e9c69d6fbb2" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.323535 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa312bfcc4c7878cb4aa4e823d7f08cb05d5351b8b33901207596e9c69d6fbb2"} err="failed to get container status \"fa312bfcc4c7878cb4aa4e823d7f08cb05d5351b8b33901207596e9c69d6fbb2\": rpc error: code = NotFound desc = could not find container \"fa312bfcc4c7878cb4aa4e823d7f08cb05d5351b8b33901207596e9c69d6fbb2\": container with ID starting with fa312bfcc4c7878cb4aa4e823d7f08cb05d5351b8b33901207596e9c69d6fbb2 not found: ID does not exist" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.330782 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb507b70-6b06-42c0-aec3-e5343f719efe-config-data\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.330854 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb507b70-6b06-42c0-aec3-e5343f719efe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.330912 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb507b70-6b06-42c0-aec3-e5343f719efe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.331077 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb507b70-6b06-42c0-aec3-e5343f719efe-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.331121 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb507b70-6b06-42c0-aec3-e5343f719efe-config-data-custom\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.331150 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb507b70-6b06-42c0-aec3-e5343f719efe-public-tls-certs\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.331195 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb507b70-6b06-42c0-aec3-e5343f719efe-scripts\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.331234 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rstnh\" (UniqueName: \"kubernetes.io/projected/eb507b70-6b06-42c0-aec3-e5343f719efe-kube-api-access-rstnh\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.331270 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb507b70-6b06-42c0-aec3-e5343f719efe-logs\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.432405 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb507b70-6b06-42c0-aec3-e5343f719efe-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.432445 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb507b70-6b06-42c0-aec3-e5343f719efe-config-data-custom\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.432469 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb507b70-6b06-42c0-aec3-e5343f719efe-public-tls-certs\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.432504 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb507b70-6b06-42c0-aec3-e5343f719efe-scripts\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.432527 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rstnh\" (UniqueName: \"kubernetes.io/projected/eb507b70-6b06-42c0-aec3-e5343f719efe-kube-api-access-rstnh\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.432550 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb507b70-6b06-42c0-aec3-e5343f719efe-logs\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.432577 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb507b70-6b06-42c0-aec3-e5343f719efe-config-data\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.432598 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb507b70-6b06-42c0-aec3-e5343f719efe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.432646 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb507b70-6b06-42c0-aec3-e5343f719efe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.433699 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb507b70-6b06-42c0-aec3-e5343f719efe-logs\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.435535 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eb507b70-6b06-42c0-aec3-e5343f719efe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.439114 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb507b70-6b06-42c0-aec3-e5343f719efe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.439120 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb507b70-6b06-42c0-aec3-e5343f719efe-public-tls-certs\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.439471 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb507b70-6b06-42c0-aec3-e5343f719efe-scripts\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.440853 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb507b70-6b06-42c0-aec3-e5343f719efe-config-data\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.443359 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb507b70-6b06-42c0-aec3-e5343f719efe-config-data-custom\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.445158 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb507b70-6b06-42c0-aec3-e5343f719efe-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.453017 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rstnh\" (UniqueName: \"kubernetes.io/projected/eb507b70-6b06-42c0-aec3-e5343f719efe-kube-api-access-rstnh\") pod \"cinder-api-0\" (UID: \"eb507b70-6b06-42c0-aec3-e5343f719efe\") " pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.604520 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 03 13:13:28 crc kubenswrapper[4578]: I1003 13:13:28.921750 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e" path="/var/lib/kubelet/pods/4aba55d1-ad9e-4803-b2ed-ff0fdd6f4f0e/volumes" Oct 03 13:13:29 crc kubenswrapper[4578]: I1003 13:13:29.026948 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 03 13:13:29 crc kubenswrapper[4578]: I1003 13:13:29.228404 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbbb1889-c411-41c1-81f2-ee4ab509acc9","Type":"ContainerStarted","Data":"56a50b2431d5a1e167b6381bf56e3a7b675abbcf7eb38b5e738174534982c3a9"} Oct 03 13:13:29 crc kubenswrapper[4578]: I1003 13:13:29.228450 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbbb1889-c411-41c1-81f2-ee4ab509acc9","Type":"ContainerStarted","Data":"533c720adf8de14e711aa8e613cb49fd93c805826ad286d5dc4b4fe8e6756c35"} Oct 03 13:13:29 crc kubenswrapper[4578]: I1003 13:13:29.231139 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"eb507b70-6b06-42c0-aec3-e5343f719efe","Type":"ContainerStarted","Data":"8c026da2ca40f2d7071b1800aedde6d18c07101f734fe4bfbfe0511ee6bbf2fd"} Oct 03 13:13:30 crc kubenswrapper[4578]: I1003 13:13:30.246207 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"eb507b70-6b06-42c0-aec3-e5343f719efe","Type":"ContainerStarted","Data":"f3ab3b715cc9aa09d8bd4ff8874e47670482c03d97a2f133a2a4a5080336398e"} Oct 03 13:13:30 crc kubenswrapper[4578]: I1003 13:13:30.247182 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"eb507b70-6b06-42c0-aec3-e5343f719efe","Type":"ContainerStarted","Data":"43398fbb6a523840dbe323833cab173c45a009a5d0cf5f4498ddb6b0a516d4f1"} Oct 03 13:13:30 crc kubenswrapper[4578]: I1003 13:13:30.247207 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 03 13:13:30 crc kubenswrapper[4578]: I1003 13:13:30.278795 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.278774467 podStartE2EDuration="2.278774467s" podCreationTimestamp="2025-10-03 13:13:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:13:30.262426068 +0000 UTC m=+1346.060898252" watchObservedRunningTime="2025-10-03 13:13:30.278774467 +0000 UTC m=+1346.077246651" Oct 03 13:13:32 crc kubenswrapper[4578]: I1003 13:13:32.828791 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:13:32 crc kubenswrapper[4578]: I1003 13:13:32.884064 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-x5hxs"] Oct 03 13:13:32 crc kubenswrapper[4578]: I1003 13:13:32.884303 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" podUID="9fdb8751-6b7b-4a75-abe6-0d5d5428cccc" containerName="dnsmasq-dns" containerID="cri-o://ef81916cf4cbd206da14a37bca646adb216da11a3d32cc1f3251c3549ad12503" gracePeriod=10 Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.067007 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.105378 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.293903 4578 generic.go:334] "Generic (PLEG): container finished" podID="9fdb8751-6b7b-4a75-abe6-0d5d5428cccc" containerID="ef81916cf4cbd206da14a37bca646adb216da11a3d32cc1f3251c3549ad12503" exitCode=0 Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.294171 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="aad29d7f-9721-491d-b3b6-5064b953ac45" containerName="cinder-scheduler" containerID="cri-o://34bfe862b66fd73388fa9e7185e5cbc50746bca0f3cc35a05193204a1a27aa5a" gracePeriod=30 Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.294526 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" event={"ID":"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc","Type":"ContainerDied","Data":"ef81916cf4cbd206da14a37bca646adb216da11a3d32cc1f3251c3549ad12503"} Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.294838 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="aad29d7f-9721-491d-b3b6-5064b953ac45" containerName="probe" containerID="cri-o://bd3d34b2b6f782764b77d72f37586480d926b7aa8b52a17b6f5abe0621841a8d" gracePeriod=30 Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.450065 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.537597 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-ovsdbserver-sb\") pod \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.537731 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-dns-swift-storage-0\") pod \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.537872 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-config\") pod \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.537891 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-ovsdbserver-nb\") pod \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.537936 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-dns-svc\") pod \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.537997 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdctb\" (UniqueName: \"kubernetes.io/projected/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-kube-api-access-sdctb\") pod \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\" (UID: \"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc\") " Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.546235 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-kube-api-access-sdctb" (OuterVolumeSpecName: "kube-api-access-sdctb") pod "9fdb8751-6b7b-4a75-abe6-0d5d5428cccc" (UID: "9fdb8751-6b7b-4a75-abe6-0d5d5428cccc"). InnerVolumeSpecName "kube-api-access-sdctb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.634926 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-64b9f57cc5-jrm9t" Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.636910 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9fdb8751-6b7b-4a75-abe6-0d5d5428cccc" (UID: "9fdb8751-6b7b-4a75-abe6-0d5d5428cccc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.637076 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9fdb8751-6b7b-4a75-abe6-0d5d5428cccc" (UID: "9fdb8751-6b7b-4a75-abe6-0d5d5428cccc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.639834 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdctb\" (UniqueName: \"kubernetes.io/projected/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-kube-api-access-sdctb\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.639863 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.639872 4578 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.649084 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-config" (OuterVolumeSpecName: "config") pod "9fdb8751-6b7b-4a75-abe6-0d5d5428cccc" (UID: "9fdb8751-6b7b-4a75-abe6-0d5d5428cccc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.660507 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9fdb8751-6b7b-4a75-abe6-0d5d5428cccc" (UID: "9fdb8751-6b7b-4a75-abe6-0d5d5428cccc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.682684 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9fdb8751-6b7b-4a75-abe6-0d5d5428cccc" (UID: "9fdb8751-6b7b-4a75-abe6-0d5d5428cccc"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.744436 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.744478 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:33 crc kubenswrapper[4578]: I1003 13:13:33.744493 4578 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:34 crc kubenswrapper[4578]: I1003 13:13:34.304988 4578 generic.go:334] "Generic (PLEG): container finished" podID="aad29d7f-9721-491d-b3b6-5064b953ac45" containerID="bd3d34b2b6f782764b77d72f37586480d926b7aa8b52a17b6f5abe0621841a8d" exitCode=0 Oct 03 13:13:34 crc kubenswrapper[4578]: I1003 13:13:34.305049 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"aad29d7f-9721-491d-b3b6-5064b953ac45","Type":"ContainerDied","Data":"bd3d34b2b6f782764b77d72f37586480d926b7aa8b52a17b6f5abe0621841a8d"} Oct 03 13:13:34 crc kubenswrapper[4578]: I1003 13:13:34.307435 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbbb1889-c411-41c1-81f2-ee4ab509acc9","Type":"ContainerStarted","Data":"f8855271a7683d1b169482c690c052dcee24bb09ad2c36500ddebe71cf0b10a7"} Oct 03 13:13:34 crc kubenswrapper[4578]: I1003 13:13:34.308985 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" event={"ID":"9fdb8751-6b7b-4a75-abe6-0d5d5428cccc","Type":"ContainerDied","Data":"d23ffb0337c0372d2785ff2edb883d34fc8ae4b387819c47f2173f0d34ec780f"} Oct 03 13:13:34 crc kubenswrapper[4578]: I1003 13:13:34.309031 4578 scope.go:117] "RemoveContainer" containerID="ef81916cf4cbd206da14a37bca646adb216da11a3d32cc1f3251c3549ad12503" Oct 03 13:13:34 crc kubenswrapper[4578]: I1003 13:13:34.309046 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-x5hxs" Oct 03 13:13:34 crc kubenswrapper[4578]: I1003 13:13:34.330574 4578 scope.go:117] "RemoveContainer" containerID="16cee472c5d94cf61cd4e5b6a1916995a6c585a2b77b67fcf84c9b254b384d44" Oct 03 13:13:34 crc kubenswrapper[4578]: I1003 13:13:34.348763 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-x5hxs"] Oct 03 13:13:34 crc kubenswrapper[4578]: I1003 13:13:34.359433 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-x5hxs"] Oct 03 13:13:34 crc kubenswrapper[4578]: I1003 13:13:34.938507 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fdb8751-6b7b-4a75-abe6-0d5d5428cccc" path="/var/lib/kubelet/pods/9fdb8751-6b7b-4a75-abe6-0d5d5428cccc/volumes" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.171052 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.270221 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-scripts\") pod \"aad29d7f-9721-491d-b3b6-5064b953ac45\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.270260 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-combined-ca-bundle\") pod \"aad29d7f-9721-491d-b3b6-5064b953ac45\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.270304 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aad29d7f-9721-491d-b3b6-5064b953ac45-etc-machine-id\") pod \"aad29d7f-9721-491d-b3b6-5064b953ac45\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.270325 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-config-data-custom\") pod \"aad29d7f-9721-491d-b3b6-5064b953ac45\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.270386 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-config-data\") pod \"aad29d7f-9721-491d-b3b6-5064b953ac45\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.270430 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5tvj\" (UniqueName: \"kubernetes.io/projected/aad29d7f-9721-491d-b3b6-5064b953ac45-kube-api-access-k5tvj\") pod \"aad29d7f-9721-491d-b3b6-5064b953ac45\" (UID: \"aad29d7f-9721-491d-b3b6-5064b953ac45\") " Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.270948 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aad29d7f-9721-491d-b3b6-5064b953ac45-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "aad29d7f-9721-491d-b3b6-5064b953ac45" (UID: "aad29d7f-9721-491d-b3b6-5064b953ac45"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.275602 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-scripts" (OuterVolumeSpecName: "scripts") pod "aad29d7f-9721-491d-b3b6-5064b953ac45" (UID: "aad29d7f-9721-491d-b3b6-5064b953ac45"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.293453 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "aad29d7f-9721-491d-b3b6-5064b953ac45" (UID: "aad29d7f-9721-491d-b3b6-5064b953ac45"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.294177 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aad29d7f-9721-491d-b3b6-5064b953ac45-kube-api-access-k5tvj" (OuterVolumeSpecName: "kube-api-access-k5tvj") pod "aad29d7f-9721-491d-b3b6-5064b953ac45" (UID: "aad29d7f-9721-491d-b3b6-5064b953ac45"). InnerVolumeSpecName "kube-api-access-k5tvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.321030 4578 generic.go:334] "Generic (PLEG): container finished" podID="aad29d7f-9721-491d-b3b6-5064b953ac45" containerID="34bfe862b66fd73388fa9e7185e5cbc50746bca0f3cc35a05193204a1a27aa5a" exitCode=0 Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.321091 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"aad29d7f-9721-491d-b3b6-5064b953ac45","Type":"ContainerDied","Data":"34bfe862b66fd73388fa9e7185e5cbc50746bca0f3cc35a05193204a1a27aa5a"} Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.321116 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"aad29d7f-9721-491d-b3b6-5064b953ac45","Type":"ContainerDied","Data":"87e8c3706d4377f934b54e5c7c1d08e05d3151a52cee54923322e13e48e518e1"} Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.321131 4578 scope.go:117] "RemoveContainer" containerID="bd3d34b2b6f782764b77d72f37586480d926b7aa8b52a17b6f5abe0621841a8d" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.321226 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.365007 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aad29d7f-9721-491d-b3b6-5064b953ac45" (UID: "aad29d7f-9721-491d-b3b6-5064b953ac45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.375147 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5tvj\" (UniqueName: \"kubernetes.io/projected/aad29d7f-9721-491d-b3b6-5064b953ac45-kube-api-access-k5tvj\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.375181 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.375190 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.375200 4578 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aad29d7f-9721-491d-b3b6-5064b953ac45-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.375208 4578 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.389811 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-config-data" (OuterVolumeSpecName: "config-data") pod "aad29d7f-9721-491d-b3b6-5064b953ac45" (UID: "aad29d7f-9721-491d-b3b6-5064b953ac45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.425521 4578 scope.go:117] "RemoveContainer" containerID="34bfe862b66fd73388fa9e7185e5cbc50746bca0f3cc35a05193204a1a27aa5a" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.477569 4578 scope.go:117] "RemoveContainer" containerID="bd3d34b2b6f782764b77d72f37586480d926b7aa8b52a17b6f5abe0621841a8d" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.478523 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad29d7f-9721-491d-b3b6-5064b953ac45-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:35 crc kubenswrapper[4578]: E1003 13:13:35.480293 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd3d34b2b6f782764b77d72f37586480d926b7aa8b52a17b6f5abe0621841a8d\": container with ID starting with bd3d34b2b6f782764b77d72f37586480d926b7aa8b52a17b6f5abe0621841a8d not found: ID does not exist" containerID="bd3d34b2b6f782764b77d72f37586480d926b7aa8b52a17b6f5abe0621841a8d" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.480327 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd3d34b2b6f782764b77d72f37586480d926b7aa8b52a17b6f5abe0621841a8d"} err="failed to get container status \"bd3d34b2b6f782764b77d72f37586480d926b7aa8b52a17b6f5abe0621841a8d\": rpc error: code = NotFound desc = could not find container \"bd3d34b2b6f782764b77d72f37586480d926b7aa8b52a17b6f5abe0621841a8d\": container with ID starting with bd3d34b2b6f782764b77d72f37586480d926b7aa8b52a17b6f5abe0621841a8d not found: ID does not exist" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.480355 4578 scope.go:117] "RemoveContainer" containerID="34bfe862b66fd73388fa9e7185e5cbc50746bca0f3cc35a05193204a1a27aa5a" Oct 03 13:13:35 crc kubenswrapper[4578]: E1003 13:13:35.480877 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34bfe862b66fd73388fa9e7185e5cbc50746bca0f3cc35a05193204a1a27aa5a\": container with ID starting with 34bfe862b66fd73388fa9e7185e5cbc50746bca0f3cc35a05193204a1a27aa5a not found: ID does not exist" containerID="34bfe862b66fd73388fa9e7185e5cbc50746bca0f3cc35a05193204a1a27aa5a" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.480942 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34bfe862b66fd73388fa9e7185e5cbc50746bca0f3cc35a05193204a1a27aa5a"} err="failed to get container status \"34bfe862b66fd73388fa9e7185e5cbc50746bca0f3cc35a05193204a1a27aa5a\": rpc error: code = NotFound desc = could not find container \"34bfe862b66fd73388fa9e7185e5cbc50746bca0f3cc35a05193204a1a27aa5a\": container with ID starting with 34bfe862b66fd73388fa9e7185e5cbc50746bca0f3cc35a05193204a1a27aa5a not found: ID does not exist" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.651697 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.656053 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.680503 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:13:35 crc kubenswrapper[4578]: E1003 13:13:35.686419 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fdb8751-6b7b-4a75-abe6-0d5d5428cccc" containerName="init" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.686461 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fdb8751-6b7b-4a75-abe6-0d5d5428cccc" containerName="init" Oct 03 13:13:35 crc kubenswrapper[4578]: E1003 13:13:35.686477 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aad29d7f-9721-491d-b3b6-5064b953ac45" containerName="cinder-scheduler" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.686485 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="aad29d7f-9721-491d-b3b6-5064b953ac45" containerName="cinder-scheduler" Oct 03 13:13:35 crc kubenswrapper[4578]: E1003 13:13:35.686497 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fdb8751-6b7b-4a75-abe6-0d5d5428cccc" containerName="dnsmasq-dns" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.686503 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fdb8751-6b7b-4a75-abe6-0d5d5428cccc" containerName="dnsmasq-dns" Oct 03 13:13:35 crc kubenswrapper[4578]: E1003 13:13:35.686544 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aad29d7f-9721-491d-b3b6-5064b953ac45" containerName="probe" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.686549 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="aad29d7f-9721-491d-b3b6-5064b953ac45" containerName="probe" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.686798 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fdb8751-6b7b-4a75-abe6-0d5d5428cccc" containerName="dnsmasq-dns" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.686814 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="aad29d7f-9721-491d-b3b6-5064b953ac45" containerName="cinder-scheduler" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.686824 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="aad29d7f-9721-491d-b3b6-5064b953ac45" containerName="probe" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.687706 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.690266 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.691963 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.800556 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c380ba54-4a4d-4e1f-8f75-4628c1f33d7c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.800607 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sx9dr\" (UniqueName: \"kubernetes.io/projected/c380ba54-4a4d-4e1f-8f75-4628c1f33d7c-kube-api-access-sx9dr\") pod \"cinder-scheduler-0\" (UID: \"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.800755 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c380ba54-4a4d-4e1f-8f75-4628c1f33d7c-scripts\") pod \"cinder-scheduler-0\" (UID: \"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.800793 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c380ba54-4a4d-4e1f-8f75-4628c1f33d7c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.800836 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c380ba54-4a4d-4e1f-8f75-4628c1f33d7c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.800908 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c380ba54-4a4d-4e1f-8f75-4628c1f33d7c-config-data\") pod \"cinder-scheduler-0\" (UID: \"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.901929 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c380ba54-4a4d-4e1f-8f75-4628c1f33d7c-config-data\") pod \"cinder-scheduler-0\" (UID: \"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.902012 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c380ba54-4a4d-4e1f-8f75-4628c1f33d7c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.902039 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sx9dr\" (UniqueName: \"kubernetes.io/projected/c380ba54-4a4d-4e1f-8f75-4628c1f33d7c-kube-api-access-sx9dr\") pod \"cinder-scheduler-0\" (UID: \"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.902073 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c380ba54-4a4d-4e1f-8f75-4628c1f33d7c-scripts\") pod \"cinder-scheduler-0\" (UID: \"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.902106 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c380ba54-4a4d-4e1f-8f75-4628c1f33d7c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.902139 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c380ba54-4a4d-4e1f-8f75-4628c1f33d7c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.902216 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c380ba54-4a4d-4e1f-8f75-4628c1f33d7c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.906808 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c380ba54-4a4d-4e1f-8f75-4628c1f33d7c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.906812 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c380ba54-4a4d-4e1f-8f75-4628c1f33d7c-config-data\") pod \"cinder-scheduler-0\" (UID: \"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.908223 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c380ba54-4a4d-4e1f-8f75-4628c1f33d7c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.909995 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c380ba54-4a4d-4e1f-8f75-4628c1f33d7c-scripts\") pod \"cinder-scheduler-0\" (UID: \"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.922921 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sx9dr\" (UniqueName: \"kubernetes.io/projected/c380ba54-4a4d-4e1f-8f75-4628c1f33d7c-kube-api-access-sx9dr\") pod \"cinder-scheduler-0\" (UID: \"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c\") " pod="openstack/cinder-scheduler-0" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.957304 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.958477 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.960435 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-z978k" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.960436 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.960564 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 03 13:13:35 crc kubenswrapper[4578]: I1003 13:13:35.975401 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.007911 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/925672c2-9609-4af9-84de-47dc5954c0dc-openstack-config\") pod \"openstackclient\" (UID: \"925672c2-9609-4af9-84de-47dc5954c0dc\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.008062 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/925672c2-9609-4af9-84de-47dc5954c0dc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"925672c2-9609-4af9-84de-47dc5954c0dc\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.008145 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/925672c2-9609-4af9-84de-47dc5954c0dc-openstack-config-secret\") pod \"openstackclient\" (UID: \"925672c2-9609-4af9-84de-47dc5954c0dc\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.008194 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5b5fx\" (UniqueName: \"kubernetes.io/projected/925672c2-9609-4af9-84de-47dc5954c0dc-kube-api-access-5b5fx\") pod \"openstackclient\" (UID: \"925672c2-9609-4af9-84de-47dc5954c0dc\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.028189 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.109161 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/925672c2-9609-4af9-84de-47dc5954c0dc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"925672c2-9609-4af9-84de-47dc5954c0dc\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.109253 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/925672c2-9609-4af9-84de-47dc5954c0dc-openstack-config-secret\") pod \"openstackclient\" (UID: \"925672c2-9609-4af9-84de-47dc5954c0dc\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.109290 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5b5fx\" (UniqueName: \"kubernetes.io/projected/925672c2-9609-4af9-84de-47dc5954c0dc-kube-api-access-5b5fx\") pod \"openstackclient\" (UID: \"925672c2-9609-4af9-84de-47dc5954c0dc\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.109315 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/925672c2-9609-4af9-84de-47dc5954c0dc-openstack-config\") pod \"openstackclient\" (UID: \"925672c2-9609-4af9-84de-47dc5954c0dc\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.110091 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/925672c2-9609-4af9-84de-47dc5954c0dc-openstack-config\") pod \"openstackclient\" (UID: \"925672c2-9609-4af9-84de-47dc5954c0dc\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.113466 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/925672c2-9609-4af9-84de-47dc5954c0dc-openstack-config-secret\") pod \"openstackclient\" (UID: \"925672c2-9609-4af9-84de-47dc5954c0dc\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.117721 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/925672c2-9609-4af9-84de-47dc5954c0dc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"925672c2-9609-4af9-84de-47dc5954c0dc\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.135498 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5b5fx\" (UniqueName: \"kubernetes.io/projected/925672c2-9609-4af9-84de-47dc5954c0dc-kube-api-access-5b5fx\") pod \"openstackclient\" (UID: \"925672c2-9609-4af9-84de-47dc5954c0dc\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.242538 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.243658 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.259213 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.266753 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.268105 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.273109 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.357435 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbbb1889-c411-41c1-81f2-ee4ab509acc9","Type":"ContainerStarted","Data":"8f2ae580dec48dd13a4379698e6ef4b6aafcb956c26934475b079bdadbb37ed8"} Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.357727 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.430502 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f409b2c2-1191-4048-b0bc-e5c449425aa8-openstack-config-secret\") pod \"openstackclient\" (UID: \"f409b2c2-1191-4048-b0bc-e5c449425aa8\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.430568 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7xt9\" (UniqueName: \"kubernetes.io/projected/f409b2c2-1191-4048-b0bc-e5c449425aa8-kube-api-access-p7xt9\") pod \"openstackclient\" (UID: \"f409b2c2-1191-4048-b0bc-e5c449425aa8\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.430665 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f409b2c2-1191-4048-b0bc-e5c449425aa8-openstack-config\") pod \"openstackclient\" (UID: \"f409b2c2-1191-4048-b0bc-e5c449425aa8\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.430786 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f409b2c2-1191-4048-b0bc-e5c449425aa8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f409b2c2-1191-4048-b0bc-e5c449425aa8\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: E1003 13:13:36.435501 4578 log.go:32] "RunPodSandbox from runtime service failed" err=< Oct 03 13:13:36 crc kubenswrapper[4578]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_925672c2-9609-4af9-84de-47dc5954c0dc_0(c308395d659046df2f63126ecdb63ada8db17bcd48646f342714304bb59f82c8): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"c308395d659046df2f63126ecdb63ada8db17bcd48646f342714304bb59f82c8" Netns:"/var/run/netns/9de9f3af-c98f-403f-97f9-0cc505349e7a" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=c308395d659046df2f63126ecdb63ada8db17bcd48646f342714304bb59f82c8;K8S_POD_UID=925672c2-9609-4af9-84de-47dc5954c0dc" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/925672c2-9609-4af9-84de-47dc5954c0dc]: expected pod UID "925672c2-9609-4af9-84de-47dc5954c0dc" but got "f409b2c2-1191-4048-b0bc-e5c449425aa8" from Kube API Oct 03 13:13:36 crc kubenswrapper[4578]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Oct 03 13:13:36 crc kubenswrapper[4578]: > Oct 03 13:13:36 crc kubenswrapper[4578]: E1003 13:13:36.435595 4578 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Oct 03 13:13:36 crc kubenswrapper[4578]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_925672c2-9609-4af9-84de-47dc5954c0dc_0(c308395d659046df2f63126ecdb63ada8db17bcd48646f342714304bb59f82c8): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"c308395d659046df2f63126ecdb63ada8db17bcd48646f342714304bb59f82c8" Netns:"/var/run/netns/9de9f3af-c98f-403f-97f9-0cc505349e7a" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=c308395d659046df2f63126ecdb63ada8db17bcd48646f342714304bb59f82c8;K8S_POD_UID=925672c2-9609-4af9-84de-47dc5954c0dc" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/925672c2-9609-4af9-84de-47dc5954c0dc]: expected pod UID "925672c2-9609-4af9-84de-47dc5954c0dc" but got "f409b2c2-1191-4048-b0bc-e5c449425aa8" from Kube API Oct 03 13:13:36 crc kubenswrapper[4578]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Oct 03 13:13:36 crc kubenswrapper[4578]: > pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.492660 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.309103978 podStartE2EDuration="10.492641956s" podCreationTimestamp="2025-10-03 13:13:26 +0000 UTC" firstStartedPulling="2025-10-03 13:13:27.153048693 +0000 UTC m=+1342.951520877" lastFinishedPulling="2025-10-03 13:13:35.336586681 +0000 UTC m=+1351.135058855" observedRunningTime="2025-10-03 13:13:36.384028071 +0000 UTC m=+1352.182500265" watchObservedRunningTime="2025-10-03 13:13:36.492641956 +0000 UTC m=+1352.291114140" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.492915 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.533050 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f409b2c2-1191-4048-b0bc-e5c449425aa8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f409b2c2-1191-4048-b0bc-e5c449425aa8\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.533722 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f409b2c2-1191-4048-b0bc-e5c449425aa8-openstack-config-secret\") pod \"openstackclient\" (UID: \"f409b2c2-1191-4048-b0bc-e5c449425aa8\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.534571 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7xt9\" (UniqueName: \"kubernetes.io/projected/f409b2c2-1191-4048-b0bc-e5c449425aa8-kube-api-access-p7xt9\") pod \"openstackclient\" (UID: \"f409b2c2-1191-4048-b0bc-e5c449425aa8\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.534735 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f409b2c2-1191-4048-b0bc-e5c449425aa8-openstack-config\") pod \"openstackclient\" (UID: \"f409b2c2-1191-4048-b0bc-e5c449425aa8\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.536284 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f409b2c2-1191-4048-b0bc-e5c449425aa8-openstack-config\") pod \"openstackclient\" (UID: \"f409b2c2-1191-4048-b0bc-e5c449425aa8\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.536755 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f409b2c2-1191-4048-b0bc-e5c449425aa8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f409b2c2-1191-4048-b0bc-e5c449425aa8\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.538490 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f409b2c2-1191-4048-b0bc-e5c449425aa8-openstack-config-secret\") pod \"openstackclient\" (UID: \"f409b2c2-1191-4048-b0bc-e5c449425aa8\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.564111 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7xt9\" (UniqueName: \"kubernetes.io/projected/f409b2c2-1191-4048-b0bc-e5c449425aa8-kube-api-access-p7xt9\") pod \"openstackclient\" (UID: \"f409b2c2-1191-4048-b0bc-e5c449425aa8\") " pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.671297 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 13:13:36 crc kubenswrapper[4578]: I1003 13:13:36.936395 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aad29d7f-9721-491d-b3b6-5064b953ac45" path="/var/lib/kubelet/pods/aad29d7f-9721-491d-b3b6-5064b953ac45/volumes" Oct 03 13:13:37 crc kubenswrapper[4578]: I1003 13:13:37.189889 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 03 13:13:37 crc kubenswrapper[4578]: I1003 13:13:37.376786 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f409b2c2-1191-4048-b0bc-e5c449425aa8","Type":"ContainerStarted","Data":"9b937bc6dd5dc018c4bf22ace2a8e418f345da7477269c34603a2d812bb96a1c"} Oct 03 13:13:37 crc kubenswrapper[4578]: I1003 13:13:37.386141 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c","Type":"ContainerStarted","Data":"027c736e8f9fcf5cea69db8dedc68c68cf8d932172558b1a2678d4c843671f29"} Oct 03 13:13:37 crc kubenswrapper[4578]: I1003 13:13:37.386177 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c","Type":"ContainerStarted","Data":"d61944912a6dad1d4dac508e28f85dae59f13419b06d3e0893c9ea00f7be2811"} Oct 03 13:13:37 crc kubenswrapper[4578]: I1003 13:13:37.386152 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 13:13:37 crc kubenswrapper[4578]: I1003 13:13:37.418568 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 13:13:37 crc kubenswrapper[4578]: I1003 13:13:37.421332 4578 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="925672c2-9609-4af9-84de-47dc5954c0dc" podUID="f409b2c2-1191-4048-b0bc-e5c449425aa8" Oct 03 13:13:37 crc kubenswrapper[4578]: I1003 13:13:37.560721 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5b5fx\" (UniqueName: \"kubernetes.io/projected/925672c2-9609-4af9-84de-47dc5954c0dc-kube-api-access-5b5fx\") pod \"925672c2-9609-4af9-84de-47dc5954c0dc\" (UID: \"925672c2-9609-4af9-84de-47dc5954c0dc\") " Oct 03 13:13:37 crc kubenswrapper[4578]: I1003 13:13:37.560836 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/925672c2-9609-4af9-84de-47dc5954c0dc-combined-ca-bundle\") pod \"925672c2-9609-4af9-84de-47dc5954c0dc\" (UID: \"925672c2-9609-4af9-84de-47dc5954c0dc\") " Oct 03 13:13:37 crc kubenswrapper[4578]: I1003 13:13:37.561025 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/925672c2-9609-4af9-84de-47dc5954c0dc-openstack-config\") pod \"925672c2-9609-4af9-84de-47dc5954c0dc\" (UID: \"925672c2-9609-4af9-84de-47dc5954c0dc\") " Oct 03 13:13:37 crc kubenswrapper[4578]: I1003 13:13:37.561054 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/925672c2-9609-4af9-84de-47dc5954c0dc-openstack-config-secret\") pod \"925672c2-9609-4af9-84de-47dc5954c0dc\" (UID: \"925672c2-9609-4af9-84de-47dc5954c0dc\") " Oct 03 13:13:37 crc kubenswrapper[4578]: I1003 13:13:37.568192 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925672c2-9609-4af9-84de-47dc5954c0dc-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "925672c2-9609-4af9-84de-47dc5954c0dc" (UID: "925672c2-9609-4af9-84de-47dc5954c0dc"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:13:37 crc kubenswrapper[4578]: I1003 13:13:37.575377 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925672c2-9609-4af9-84de-47dc5954c0dc-kube-api-access-5b5fx" (OuterVolumeSpecName: "kube-api-access-5b5fx") pod "925672c2-9609-4af9-84de-47dc5954c0dc" (UID: "925672c2-9609-4af9-84de-47dc5954c0dc"). InnerVolumeSpecName "kube-api-access-5b5fx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:13:37 crc kubenswrapper[4578]: I1003 13:13:37.575591 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925672c2-9609-4af9-84de-47dc5954c0dc-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "925672c2-9609-4af9-84de-47dc5954c0dc" (UID: "925672c2-9609-4af9-84de-47dc5954c0dc"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:37 crc kubenswrapper[4578]: I1003 13:13:37.576546 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925672c2-9609-4af9-84de-47dc5954c0dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "925672c2-9609-4af9-84de-47dc5954c0dc" (UID: "925672c2-9609-4af9-84de-47dc5954c0dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:37 crc kubenswrapper[4578]: I1003 13:13:37.665032 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5b5fx\" (UniqueName: \"kubernetes.io/projected/925672c2-9609-4af9-84de-47dc5954c0dc-kube-api-access-5b5fx\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:37 crc kubenswrapper[4578]: I1003 13:13:37.665089 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/925672c2-9609-4af9-84de-47dc5954c0dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:37 crc kubenswrapper[4578]: I1003 13:13:37.665101 4578 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/925672c2-9609-4af9-84de-47dc5954c0dc-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:37 crc kubenswrapper[4578]: I1003 13:13:37.665113 4578 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/925672c2-9609-4af9-84de-47dc5954c0dc-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:38 crc kubenswrapper[4578]: I1003 13:13:38.398525 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 03 13:13:38 crc kubenswrapper[4578]: I1003 13:13:38.412418 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c380ba54-4a4d-4e1f-8f75-4628c1f33d7c","Type":"ContainerStarted","Data":"d0c1982f326201fba3b5f48c8661b333de68b49034f5548a4b673cc5119fbc6e"} Oct 03 13:13:38 crc kubenswrapper[4578]: I1003 13:13:38.436606 4578 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="925672c2-9609-4af9-84de-47dc5954c0dc" podUID="f409b2c2-1191-4048-b0bc-e5c449425aa8" Oct 03 13:13:38 crc kubenswrapper[4578]: I1003 13:13:38.440106 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.440076673 podStartE2EDuration="3.440076673s" podCreationTimestamp="2025-10-03 13:13:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:13:38.43004311 +0000 UTC m=+1354.228515294" watchObservedRunningTime="2025-10-03 13:13:38.440076673 +0000 UTC m=+1354.238548857" Oct 03 13:13:38 crc kubenswrapper[4578]: I1003 13:13:38.922776 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925672c2-9609-4af9-84de-47dc5954c0dc" path="/var/lib/kubelet/pods/925672c2-9609-4af9-84de-47dc5954c0dc/volumes" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.367314 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-57dc876bc5-gvzk7"] Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.369217 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.372666 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.372684 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.372953 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.399679 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-57dc876bc5-gvzk7"] Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.529036 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-combined-ca-bundle\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.529104 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-config-data\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.529171 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-run-httpd\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.529196 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-log-httpd\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.529218 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-public-tls-certs\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.529291 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjrqk\" (UniqueName: \"kubernetes.io/projected/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-kube-api-access-mjrqk\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.529327 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-internal-tls-certs\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.529370 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-etc-swift\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.630746 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjrqk\" (UniqueName: \"kubernetes.io/projected/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-kube-api-access-mjrqk\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.630806 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-internal-tls-certs\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.630864 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-etc-swift\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.630969 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-combined-ca-bundle\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.631012 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-config-data\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.631076 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-run-httpd\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.631106 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-log-httpd\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.631128 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-public-tls-certs\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.633165 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-run-httpd\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.636368 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-log-httpd\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.639203 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-config-data\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.640731 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-internal-tls-certs\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.644158 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-combined-ca-bundle\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.647509 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-public-tls-certs\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.653266 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-etc-swift\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.666364 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjrqk\" (UniqueName: \"kubernetes.io/projected/bd42e3ab-aa85-453e-abfd-7b6aa55e9674-kube-api-access-mjrqk\") pod \"swift-proxy-57dc876bc5-gvzk7\" (UID: \"bd42e3ab-aa85-453e-abfd-7b6aa55e9674\") " pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:40 crc kubenswrapper[4578]: I1003 13:13:40.696869 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:41 crc kubenswrapper[4578]: I1003 13:13:41.028728 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 03 13:13:41 crc kubenswrapper[4578]: I1003 13:13:41.224540 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 03 13:13:41 crc kubenswrapper[4578]: I1003 13:13:41.362225 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-57dc876bc5-gvzk7"] Oct 03 13:13:41 crc kubenswrapper[4578]: I1003 13:13:41.445081 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-57dc876bc5-gvzk7" event={"ID":"bd42e3ab-aa85-453e-abfd-7b6aa55e9674","Type":"ContainerStarted","Data":"259ab75b2b9e71eafd029cce0cc549c840e48c29635534bea721c5cc7e79407a"} Oct 03 13:13:42 crc kubenswrapper[4578]: I1003 13:13:42.166494 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:13:42 crc kubenswrapper[4578]: I1003 13:13:42.167131 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" containerName="ceilometer-central-agent" containerID="cri-o://533c720adf8de14e711aa8e613cb49fd93c805826ad286d5dc4b4fe8e6756c35" gracePeriod=30 Oct 03 13:13:42 crc kubenswrapper[4578]: I1003 13:13:42.167267 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" containerName="proxy-httpd" containerID="cri-o://8f2ae580dec48dd13a4379698e6ef4b6aafcb956c26934475b079bdadbb37ed8" gracePeriod=30 Oct 03 13:13:42 crc kubenswrapper[4578]: I1003 13:13:42.167322 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" containerName="sg-core" containerID="cri-o://f8855271a7683d1b169482c690c052dcee24bb09ad2c36500ddebe71cf0b10a7" gracePeriod=30 Oct 03 13:13:42 crc kubenswrapper[4578]: I1003 13:13:42.167376 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" containerName="ceilometer-notification-agent" containerID="cri-o://56a50b2431d5a1e167b6381bf56e3a7b675abbcf7eb38b5e738174534982c3a9" gracePeriod=30 Oct 03 13:13:42 crc kubenswrapper[4578]: I1003 13:13:42.458114 4578 generic.go:334] "Generic (PLEG): container finished" podID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" containerID="8f2ae580dec48dd13a4379698e6ef4b6aafcb956c26934475b079bdadbb37ed8" exitCode=0 Oct 03 13:13:42 crc kubenswrapper[4578]: I1003 13:13:42.458151 4578 generic.go:334] "Generic (PLEG): container finished" podID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" containerID="f8855271a7683d1b169482c690c052dcee24bb09ad2c36500ddebe71cf0b10a7" exitCode=2 Oct 03 13:13:42 crc kubenswrapper[4578]: I1003 13:13:42.458192 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbbb1889-c411-41c1-81f2-ee4ab509acc9","Type":"ContainerDied","Data":"8f2ae580dec48dd13a4379698e6ef4b6aafcb956c26934475b079bdadbb37ed8"} Oct 03 13:13:42 crc kubenswrapper[4578]: I1003 13:13:42.458253 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbbb1889-c411-41c1-81f2-ee4ab509acc9","Type":"ContainerDied","Data":"f8855271a7683d1b169482c690c052dcee24bb09ad2c36500ddebe71cf0b10a7"} Oct 03 13:13:42 crc kubenswrapper[4578]: I1003 13:13:42.461348 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-57dc876bc5-gvzk7" event={"ID":"bd42e3ab-aa85-453e-abfd-7b6aa55e9674","Type":"ContainerStarted","Data":"45aff087a3e117b74042a218600e8186f3d68732576cfef1f87cb751550e8406"} Oct 03 13:13:42 crc kubenswrapper[4578]: I1003 13:13:42.461901 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:42 crc kubenswrapper[4578]: I1003 13:13:42.461915 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-57dc876bc5-gvzk7" event={"ID":"bd42e3ab-aa85-453e-abfd-7b6aa55e9674","Type":"ContainerStarted","Data":"bb7f172676862c2cf15dff9865a797d312175c10f89d56a5d1a407d5d60d14c0"} Oct 03 13:13:42 crc kubenswrapper[4578]: I1003 13:13:42.482994 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-57dc876bc5-gvzk7" podStartSLOduration=2.482969589 podStartE2EDuration="2.482969589s" podCreationTimestamp="2025-10-03 13:13:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:13:42.480897164 +0000 UTC m=+1358.279369358" watchObservedRunningTime="2025-10-03 13:13:42.482969589 +0000 UTC m=+1358.281441783" Oct 03 13:13:43 crc kubenswrapper[4578]: I1003 13:13:43.476733 4578 generic.go:334] "Generic (PLEG): container finished" podID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" containerID="533c720adf8de14e711aa8e613cb49fd93c805826ad286d5dc4b4fe8e6756c35" exitCode=0 Oct 03 13:13:43 crc kubenswrapper[4578]: I1003 13:13:43.476807 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbbb1889-c411-41c1-81f2-ee4ab509acc9","Type":"ContainerDied","Data":"533c720adf8de14e711aa8e613cb49fd93c805826ad286d5dc4b4fe8e6756c35"} Oct 03 13:13:43 crc kubenswrapper[4578]: I1003 13:13:43.477024 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:46 crc kubenswrapper[4578]: I1003 13:13:46.602457 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 03 13:13:47 crc kubenswrapper[4578]: I1003 13:13:47.558881 4578 generic.go:334] "Generic (PLEG): container finished" podID="d49d623e-2cd6-4786-b1da-34b78b89a134" containerID="043ba4870e4db2272214244e538587f588862b6e71b023c4ea7dfa6638da57bd" exitCode=137 Oct 03 13:13:47 crc kubenswrapper[4578]: I1003 13:13:47.558925 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68684899bd-2cwll" event={"ID":"d49d623e-2cd6-4786-b1da-34b78b89a134","Type":"ContainerDied","Data":"043ba4870e4db2272214244e538587f588862b6e71b023c4ea7dfa6638da57bd"} Oct 03 13:13:47 crc kubenswrapper[4578]: I1003 13:13:47.565954 4578 generic.go:334] "Generic (PLEG): container finished" podID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" containerID="56a50b2431d5a1e167b6381bf56e3a7b675abbcf7eb38b5e738174534982c3a9" exitCode=0 Oct 03 13:13:47 crc kubenswrapper[4578]: I1003 13:13:47.566031 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbbb1889-c411-41c1-81f2-ee4ab509acc9","Type":"ContainerDied","Data":"56a50b2431d5a1e167b6381bf56e3a7b675abbcf7eb38b5e738174534982c3a9"} Oct 03 13:13:47 crc kubenswrapper[4578]: I1003 13:13:47.569033 4578 generic.go:334] "Generic (PLEG): container finished" podID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerID="033a127d7b3dc6a5469b25fc2682629ad13d3e1ed8dd904aad8edc6112ea940b" exitCode=137 Oct 03 13:13:47 crc kubenswrapper[4578]: I1003 13:13:47.569081 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b58cf9568-f8nng" event={"ID":"a54aea6a-c051-4832-b764-3ef53a9cc3a1","Type":"ContainerDied","Data":"033a127d7b3dc6a5469b25fc2682629ad13d3e1ed8dd904aad8edc6112ea940b"} Oct 03 13:13:50 crc kubenswrapper[4578]: I1003 13:13:50.711481 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:50 crc kubenswrapper[4578]: I1003 13:13:50.714509 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-57dc876bc5-gvzk7" Oct 03 13:13:51 crc kubenswrapper[4578]: I1003 13:13:51.855910 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:13:51 crc kubenswrapper[4578]: I1003 13:13:51.881625 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbbb1889-c411-41c1-81f2-ee4ab509acc9-run-httpd\") pod \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " Oct 03 13:13:51 crc kubenswrapper[4578]: I1003 13:13:51.881723 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-scripts\") pod \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " Oct 03 13:13:51 crc kubenswrapper[4578]: I1003 13:13:51.881760 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-combined-ca-bundle\") pod \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " Oct 03 13:13:51 crc kubenswrapper[4578]: I1003 13:13:51.881821 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-sg-core-conf-yaml\") pod \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " Oct 03 13:13:51 crc kubenswrapper[4578]: I1003 13:13:51.881861 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-config-data\") pod \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " Oct 03 13:13:51 crc kubenswrapper[4578]: I1003 13:13:51.881885 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbbb1889-c411-41c1-81f2-ee4ab509acc9-log-httpd\") pod \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " Oct 03 13:13:51 crc kubenswrapper[4578]: I1003 13:13:51.881914 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdccd\" (UniqueName: \"kubernetes.io/projected/dbbb1889-c411-41c1-81f2-ee4ab509acc9-kube-api-access-rdccd\") pod \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\" (UID: \"dbbb1889-c411-41c1-81f2-ee4ab509acc9\") " Oct 03 13:13:51 crc kubenswrapper[4578]: I1003 13:13:51.895067 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbbb1889-c411-41c1-81f2-ee4ab509acc9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "dbbb1889-c411-41c1-81f2-ee4ab509acc9" (UID: "dbbb1889-c411-41c1-81f2-ee4ab509acc9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:13:51 crc kubenswrapper[4578]: I1003 13:13:51.895526 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbbb1889-c411-41c1-81f2-ee4ab509acc9-kube-api-access-rdccd" (OuterVolumeSpecName: "kube-api-access-rdccd") pod "dbbb1889-c411-41c1-81f2-ee4ab509acc9" (UID: "dbbb1889-c411-41c1-81f2-ee4ab509acc9"). InnerVolumeSpecName "kube-api-access-rdccd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:13:51 crc kubenswrapper[4578]: I1003 13:13:51.899591 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbbb1889-c411-41c1-81f2-ee4ab509acc9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "dbbb1889-c411-41c1-81f2-ee4ab509acc9" (UID: "dbbb1889-c411-41c1-81f2-ee4ab509acc9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:13:51 crc kubenswrapper[4578]: I1003 13:13:51.914364 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-scripts" (OuterVolumeSpecName: "scripts") pod "dbbb1889-c411-41c1-81f2-ee4ab509acc9" (UID: "dbbb1889-c411-41c1-81f2-ee4ab509acc9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:51 crc kubenswrapper[4578]: I1003 13:13:51.963484 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "dbbb1889-c411-41c1-81f2-ee4ab509acc9" (UID: "dbbb1889-c411-41c1-81f2-ee4ab509acc9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:51 crc kubenswrapper[4578]: I1003 13:13:51.982994 4578 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:51 crc kubenswrapper[4578]: I1003 13:13:51.983022 4578 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbbb1889-c411-41c1-81f2-ee4ab509acc9-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:51 crc kubenswrapper[4578]: I1003 13:13:51.983031 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdccd\" (UniqueName: \"kubernetes.io/projected/dbbb1889-c411-41c1-81f2-ee4ab509acc9-kube-api-access-rdccd\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:51 crc kubenswrapper[4578]: I1003 13:13:51.983039 4578 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbbb1889-c411-41c1-81f2-ee4ab509acc9-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:51 crc kubenswrapper[4578]: I1003 13:13:51.983049 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.002999 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dbbb1889-c411-41c1-81f2-ee4ab509acc9" (UID: "dbbb1889-c411-41c1-81f2-ee4ab509acc9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.044174 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-config-data" (OuterVolumeSpecName: "config-data") pod "dbbb1889-c411-41c1-81f2-ee4ab509acc9" (UID: "dbbb1889-c411-41c1-81f2-ee4ab509acc9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.084226 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.084257 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbbb1889-c411-41c1-81f2-ee4ab509acc9-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.460559 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.461138 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ec6cb50d-e8ea-450c-ba3c-e718e03c41a3" containerName="glance-log" containerID="cri-o://44e11f2ae23fddbb737f7b00fdf3e82e831289bd9ebbaf9260098bbe8900732b" gracePeriod=30 Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.461271 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ec6cb50d-e8ea-450c-ba3c-e718e03c41a3" containerName="glance-httpd" containerID="cri-o://930e4b153894449384f40e1fcaa46f1750f077197fa07f10a8b4af27ab85795a" gracePeriod=30 Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.630941 4578 generic.go:334] "Generic (PLEG): container finished" podID="ec6cb50d-e8ea-450c-ba3c-e718e03c41a3" containerID="44e11f2ae23fddbb737f7b00fdf3e82e831289bd9ebbaf9260098bbe8900732b" exitCode=143 Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.631013 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3","Type":"ContainerDied","Data":"44e11f2ae23fddbb737f7b00fdf3e82e831289bd9ebbaf9260098bbe8900732b"} Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.634704 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68684899bd-2cwll" event={"ID":"d49d623e-2cd6-4786-b1da-34b78b89a134","Type":"ContainerStarted","Data":"383aa1a23a7379fea51b3db0176d20c04f2cc353b1845fbbace3a5521aa9a565"} Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.637564 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b58cf9568-f8nng" event={"ID":"a54aea6a-c051-4832-b764-3ef53a9cc3a1","Type":"ContainerStarted","Data":"de8222a725afcbc467d7921332c4ab3214da8e901c323972f2737910bf0d32c3"} Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.641456 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbbb1889-c411-41c1-81f2-ee4ab509acc9","Type":"ContainerDied","Data":"38bb1d9921b1f9d83a95bcfbab52467ba6b8287e14cdd572fc44e923def3821d"} Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.641513 4578 scope.go:117] "RemoveContainer" containerID="8f2ae580dec48dd13a4379698e6ef4b6aafcb956c26934475b079bdadbb37ed8" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.641752 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.646786 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f409b2c2-1191-4048-b0bc-e5c449425aa8","Type":"ContainerStarted","Data":"7c4f8a217e60e61f3a9df921d0d475dd223405e6f54c1962a6bd43cee967ce57"} Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.664335 4578 scope.go:117] "RemoveContainer" containerID="f8855271a7683d1b169482c690c052dcee24bb09ad2c36500ddebe71cf0b10a7" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.687448 4578 scope.go:117] "RemoveContainer" containerID="56a50b2431d5a1e167b6381bf56e3a7b675abbcf7eb38b5e738174534982c3a9" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.688549 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.710211 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.731833 4578 scope.go:117] "RemoveContainer" containerID="533c720adf8de14e711aa8e613cb49fd93c805826ad286d5dc4b4fe8e6756c35" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.734032 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:13:52 crc kubenswrapper[4578]: E1003 13:13:52.734478 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" containerName="ceilometer-notification-agent" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.734503 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" containerName="ceilometer-notification-agent" Oct 03 13:13:52 crc kubenswrapper[4578]: E1003 13:13:52.734527 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" containerName="ceilometer-central-agent" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.734537 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" containerName="ceilometer-central-agent" Oct 03 13:13:52 crc kubenswrapper[4578]: E1003 13:13:52.734556 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" containerName="sg-core" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.734563 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" containerName="sg-core" Oct 03 13:13:52 crc kubenswrapper[4578]: E1003 13:13:52.734581 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" containerName="proxy-httpd" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.734590 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" containerName="proxy-httpd" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.734833 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" containerName="ceilometer-notification-agent" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.734862 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" containerName="ceilometer-central-agent" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.734876 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" containerName="proxy-httpd" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.734892 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" containerName="sg-core" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.736897 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.740439 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.405787494 podStartE2EDuration="16.740424224s" podCreationTimestamp="2025-10-03 13:13:36 +0000 UTC" firstStartedPulling="2025-10-03 13:13:37.199350389 +0000 UTC m=+1352.997822563" lastFinishedPulling="2025-10-03 13:13:51.533987109 +0000 UTC m=+1367.332459293" observedRunningTime="2025-10-03 13:13:52.734601603 +0000 UTC m=+1368.533073797" watchObservedRunningTime="2025-10-03 13:13:52.740424224 +0000 UTC m=+1368.538896408" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.741072 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.747869 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.769473 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.796591 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdwd4\" (UniqueName: \"kubernetes.io/projected/a925d503-688a-4a63-a6d5-1533a9a22aac-kube-api-access-qdwd4\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.796666 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-config-data\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.796801 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a925d503-688a-4a63-a6d5-1533a9a22aac-run-httpd\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.796877 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.796917 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-scripts\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.796947 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a925d503-688a-4a63-a6d5-1533a9a22aac-log-httpd\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.796982 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.898203 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a925d503-688a-4a63-a6d5-1533a9a22aac-run-httpd\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.898296 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.898332 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-scripts\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.898366 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a925d503-688a-4a63-a6d5-1533a9a22aac-log-httpd\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.898390 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.898469 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdwd4\" (UniqueName: \"kubernetes.io/projected/a925d503-688a-4a63-a6d5-1533a9a22aac-kube-api-access-qdwd4\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.898513 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-config-data\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.898659 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a925d503-688a-4a63-a6d5-1533a9a22aac-run-httpd\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.899972 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a925d503-688a-4a63-a6d5-1533a9a22aac-log-httpd\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.906554 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-config-data\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.917115 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.917750 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-scripts\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.918075 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.922235 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbbb1889-c411-41c1-81f2-ee4ab509acc9" path="/var/lib/kubelet/pods/dbbb1889-c411-41c1-81f2-ee4ab509acc9/volumes" Oct 03 13:13:52 crc kubenswrapper[4578]: I1003 13:13:52.927432 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdwd4\" (UniqueName: \"kubernetes.io/projected/a925d503-688a-4a63-a6d5-1533a9a22aac-kube-api-access-qdwd4\") pod \"ceilometer-0\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " pod="openstack/ceilometer-0" Oct 03 13:13:53 crc kubenswrapper[4578]: I1003 13:13:53.089859 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:13:53 crc kubenswrapper[4578]: I1003 13:13:53.623010 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:13:53 crc kubenswrapper[4578]: I1003 13:13:53.692391 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a925d503-688a-4a63-a6d5-1533a9a22aac","Type":"ContainerStarted","Data":"dcbaf256bf4768b662fb7ce87eaa185bef1264b7eb65a59cc2eeeba65ba377f3"} Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.479704 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-sdwwm"] Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.482267 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sdwwm" Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.497833 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-sdwwm"] Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.630583 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcwpf\" (UniqueName: \"kubernetes.io/projected/5533c61e-c654-4d79-9431-88bc37cff6d8-kube-api-access-lcwpf\") pod \"nova-api-db-create-sdwwm\" (UID: \"5533c61e-c654-4d79-9431-88bc37cff6d8\") " pod="openstack/nova-api-db-create-sdwwm" Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.659606 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-56phj"] Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.661322 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-56phj" Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.679549 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.679850 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="25ef994a-0092-4e72-a636-f51bbd1a8a99" containerName="glance-log" containerID="cri-o://666b5fd3073ef524529191139ec8cb50cfbb3d1d8b1a4b19cea1ea993e443544" gracePeriod=30 Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.679995 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="25ef994a-0092-4e72-a636-f51bbd1a8a99" containerName="glance-httpd" containerID="cri-o://2323557f12cf7f4bfe4a74ca5af7e4ce03b61bb618e398e5eba8cfa1c269bcda" gracePeriod=30 Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.702968 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-56phj"] Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.720009 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a925d503-688a-4a63-a6d5-1533a9a22aac","Type":"ContainerStarted","Data":"8e9ae4feb2c38fbae277c3dbbdc74ea68b47414938aebd6eabba698f4914b621"} Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.732396 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcwpf\" (UniqueName: \"kubernetes.io/projected/5533c61e-c654-4d79-9431-88bc37cff6d8-kube-api-access-lcwpf\") pod \"nova-api-db-create-sdwwm\" (UID: \"5533c61e-c654-4d79-9431-88bc37cff6d8\") " pod="openstack/nova-api-db-create-sdwwm" Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.764592 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcwpf\" (UniqueName: \"kubernetes.io/projected/5533c61e-c654-4d79-9431-88bc37cff6d8-kube-api-access-lcwpf\") pod \"nova-api-db-create-sdwwm\" (UID: \"5533c61e-c654-4d79-9431-88bc37cff6d8\") " pod="openstack/nova-api-db-create-sdwwm" Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.778759 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-4t996"] Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.780123 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4t996" Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.824698 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-4t996"] Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.844128 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-459br\" (UniqueName: \"kubernetes.io/projected/cafa819d-515a-4e7a-8f97-e97522c4b31c-kube-api-access-459br\") pod \"nova-cell0-db-create-56phj\" (UID: \"cafa819d-515a-4e7a-8f97-e97522c4b31c\") " pod="openstack/nova-cell0-db-create-56phj" Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.850610 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sdwwm" Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.945849 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58nkm\" (UniqueName: \"kubernetes.io/projected/9f94584d-3f52-4dcc-9c36-7f1174df9753-kube-api-access-58nkm\") pod \"nova-cell1-db-create-4t996\" (UID: \"9f94584d-3f52-4dcc-9c36-7f1174df9753\") " pod="openstack/nova-cell1-db-create-4t996" Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.945970 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-459br\" (UniqueName: \"kubernetes.io/projected/cafa819d-515a-4e7a-8f97-e97522c4b31c-kube-api-access-459br\") pod \"nova-cell0-db-create-56phj\" (UID: \"cafa819d-515a-4e7a-8f97-e97522c4b31c\") " pod="openstack/nova-cell0-db-create-56phj" Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.969212 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-459br\" (UniqueName: \"kubernetes.io/projected/cafa819d-515a-4e7a-8f97-e97522c4b31c-kube-api-access-459br\") pod \"nova-cell0-db-create-56phj\" (UID: \"cafa819d-515a-4e7a-8f97-e97522c4b31c\") " pod="openstack/nova-cell0-db-create-56phj" Oct 03 13:13:54 crc kubenswrapper[4578]: I1003 13:13:54.979852 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-56phj" Oct 03 13:13:55 crc kubenswrapper[4578]: I1003 13:13:55.047444 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58nkm\" (UniqueName: \"kubernetes.io/projected/9f94584d-3f52-4dcc-9c36-7f1174df9753-kube-api-access-58nkm\") pod \"nova-cell1-db-create-4t996\" (UID: \"9f94584d-3f52-4dcc-9c36-7f1174df9753\") " pod="openstack/nova-cell1-db-create-4t996" Oct 03 13:13:55 crc kubenswrapper[4578]: I1003 13:13:55.092709 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58nkm\" (UniqueName: \"kubernetes.io/projected/9f94584d-3f52-4dcc-9c36-7f1174df9753-kube-api-access-58nkm\") pod \"nova-cell1-db-create-4t996\" (UID: \"9f94584d-3f52-4dcc-9c36-7f1174df9753\") " pod="openstack/nova-cell1-db-create-4t996" Oct 03 13:13:55 crc kubenswrapper[4578]: I1003 13:13:55.178445 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4t996" Oct 03 13:13:55 crc kubenswrapper[4578]: I1003 13:13:55.389227 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-sdwwm"] Oct 03 13:13:55 crc kubenswrapper[4578]: I1003 13:13:55.568254 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-56phj"] Oct 03 13:13:55 crc kubenswrapper[4578]: I1003 13:13:55.744171 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-sdwwm" event={"ID":"5533c61e-c654-4d79-9431-88bc37cff6d8","Type":"ContainerStarted","Data":"7475f482cf997546210082ce0180d9e0ab5ba20c5af3520d84f2331a16c340fa"} Oct 03 13:13:55 crc kubenswrapper[4578]: I1003 13:13:55.761602 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-56phj" event={"ID":"cafa819d-515a-4e7a-8f97-e97522c4b31c","Type":"ContainerStarted","Data":"fa9ebff312d6df85772f1d9c2b2574895c5445f2785be9dbe0207f1eb6a06923"} Oct 03 13:13:55 crc kubenswrapper[4578]: I1003 13:13:55.780928 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a925d503-688a-4a63-a6d5-1533a9a22aac","Type":"ContainerStarted","Data":"6b5fa4fe8900e80130f7c569ae25f05762f0b391172a3db23a44840feff5ad7d"} Oct 03 13:13:55 crc kubenswrapper[4578]: I1003 13:13:55.807765 4578 generic.go:334] "Generic (PLEG): container finished" podID="25ef994a-0092-4e72-a636-f51bbd1a8a99" containerID="666b5fd3073ef524529191139ec8cb50cfbb3d1d8b1a4b19cea1ea993e443544" exitCode=143 Oct 03 13:13:55 crc kubenswrapper[4578]: I1003 13:13:55.808024 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"25ef994a-0092-4e72-a636-f51bbd1a8a99","Type":"ContainerDied","Data":"666b5fd3073ef524529191139ec8cb50cfbb3d1d8b1a4b19cea1ea993e443544"} Oct 03 13:13:55 crc kubenswrapper[4578]: I1003 13:13:55.822312 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-4t996"] Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.422504 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.580492 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.581878 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-scripts\") pod \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.582142 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-public-tls-certs\") pod \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.582292 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-httpd-run\") pod \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.582414 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-config-data\") pod \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.582565 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-logs\") pod \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.582723 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-combined-ca-bundle\") pod \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.582881 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7s8bm\" (UniqueName: \"kubernetes.io/projected/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-kube-api-access-7s8bm\") pod \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\" (UID: \"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3\") " Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.583531 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ec6cb50d-e8ea-450c-ba3c-e718e03c41a3" (UID: "ec6cb50d-e8ea-450c-ba3c-e718e03c41a3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.583826 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-logs" (OuterVolumeSpecName: "logs") pod "ec6cb50d-e8ea-450c-ba3c-e718e03c41a3" (UID: "ec6cb50d-e8ea-450c-ba3c-e718e03c41a3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.584618 4578 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.584781 4578 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.606447 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-scripts" (OuterVolumeSpecName: "scripts") pod "ec6cb50d-e8ea-450c-ba3c-e718e03c41a3" (UID: "ec6cb50d-e8ea-450c-ba3c-e718e03c41a3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.622493 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "ec6cb50d-e8ea-450c-ba3c-e718e03c41a3" (UID: "ec6cb50d-e8ea-450c-ba3c-e718e03c41a3"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.622495 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-kube-api-access-7s8bm" (OuterVolumeSpecName: "kube-api-access-7s8bm") pod "ec6cb50d-e8ea-450c-ba3c-e718e03c41a3" (UID: "ec6cb50d-e8ea-450c-ba3c-e718e03c41a3"). InnerVolumeSpecName "kube-api-access-7s8bm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.663835 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec6cb50d-e8ea-450c-ba3c-e718e03c41a3" (UID: "ec6cb50d-e8ea-450c-ba3c-e718e03c41a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.686394 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7s8bm\" (UniqueName: \"kubernetes.io/projected/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-kube-api-access-7s8bm\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.686450 4578 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.686462 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.686472 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.705915 4578 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.713759 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-config-data" (OuterVolumeSpecName: "config-data") pod "ec6cb50d-e8ea-450c-ba3c-e718e03c41a3" (UID: "ec6cb50d-e8ea-450c-ba3c-e718e03c41a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.736838 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ec6cb50d-e8ea-450c-ba3c-e718e03c41a3" (UID: "ec6cb50d-e8ea-450c-ba3c-e718e03c41a3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.780848 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.780885 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.788214 4578 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.788240 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.788250 4578 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.818557 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a925d503-688a-4a63-a6d5-1533a9a22aac","Type":"ContainerStarted","Data":"6853ac4b339f037dbd2c66955d65aaca7d497983ca6a0d31ff77dc5e57434c4e"} Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.819962 4578 generic.go:334] "Generic (PLEG): container finished" podID="9f94584d-3f52-4dcc-9c36-7f1174df9753" containerID="0381801bf87524ccc1641c4cf296156a2ee3996c771b45e4e60575214a418d27" exitCode=0 Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.820025 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4t996" event={"ID":"9f94584d-3f52-4dcc-9c36-7f1174df9753","Type":"ContainerDied","Data":"0381801bf87524ccc1641c4cf296156a2ee3996c771b45e4e60575214a418d27"} Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.820056 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4t996" event={"ID":"9f94584d-3f52-4dcc-9c36-7f1174df9753","Type":"ContainerStarted","Data":"45eaf9479201e417fa9605615adcbaaf3f43673c6dc34090273e7df06bc89649"} Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.823662 4578 generic.go:334] "Generic (PLEG): container finished" podID="5533c61e-c654-4d79-9431-88bc37cff6d8" containerID="439f5b895a77cc3be3500cced64bdf342ce57baa36cd07c7c3e9563b9cacefce" exitCode=0 Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.823720 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-sdwwm" event={"ID":"5533c61e-c654-4d79-9431-88bc37cff6d8","Type":"ContainerDied","Data":"439f5b895a77cc3be3500cced64bdf342ce57baa36cd07c7c3e9563b9cacefce"} Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.829435 4578 generic.go:334] "Generic (PLEG): container finished" podID="ec6cb50d-e8ea-450c-ba3c-e718e03c41a3" containerID="930e4b153894449384f40e1fcaa46f1750f077197fa07f10a8b4af27ab85795a" exitCode=0 Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.829502 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3","Type":"ContainerDied","Data":"930e4b153894449384f40e1fcaa46f1750f077197fa07f10a8b4af27ab85795a"} Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.829525 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ec6cb50d-e8ea-450c-ba3c-e718e03c41a3","Type":"ContainerDied","Data":"c84cdde022f3813331f17ad5d123c3f66050af65f2e5bc51d7a22a5c7a232e82"} Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.829542 4578 scope.go:117] "RemoveContainer" containerID="930e4b153894449384f40e1fcaa46f1750f077197fa07f10a8b4af27ab85795a" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.830025 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.848806 4578 generic.go:334] "Generic (PLEG): container finished" podID="cafa819d-515a-4e7a-8f97-e97522c4b31c" containerID="6f40a44541545d416250934fd25814d456a7fe570a45d77ca7f4b22f8992f19d" exitCode=0 Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.848858 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-56phj" event={"ID":"cafa819d-515a-4e7a-8f97-e97522c4b31c","Type":"ContainerDied","Data":"6f40a44541545d416250934fd25814d456a7fe570a45d77ca7f4b22f8992f19d"} Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.904861 4578 scope.go:117] "RemoveContainer" containerID="44e11f2ae23fddbb737f7b00fdf3e82e831289bd9ebbaf9260098bbe8900732b" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.947399 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.949792 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.949886 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.959224 4578 scope.go:117] "RemoveContainer" containerID="930e4b153894449384f40e1fcaa46f1750f077197fa07f10a8b4af27ab85795a" Oct 03 13:13:56 crc kubenswrapper[4578]: E1003 13:13:56.960588 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"930e4b153894449384f40e1fcaa46f1750f077197fa07f10a8b4af27ab85795a\": container with ID starting with 930e4b153894449384f40e1fcaa46f1750f077197fa07f10a8b4af27ab85795a not found: ID does not exist" containerID="930e4b153894449384f40e1fcaa46f1750f077197fa07f10a8b4af27ab85795a" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.960657 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"930e4b153894449384f40e1fcaa46f1750f077197fa07f10a8b4af27ab85795a"} err="failed to get container status \"930e4b153894449384f40e1fcaa46f1750f077197fa07f10a8b4af27ab85795a\": rpc error: code = NotFound desc = could not find container \"930e4b153894449384f40e1fcaa46f1750f077197fa07f10a8b4af27ab85795a\": container with ID starting with 930e4b153894449384f40e1fcaa46f1750f077197fa07f10a8b4af27ab85795a not found: ID does not exist" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.960690 4578 scope.go:117] "RemoveContainer" containerID="44e11f2ae23fddbb737f7b00fdf3e82e831289bd9ebbaf9260098bbe8900732b" Oct 03 13:13:56 crc kubenswrapper[4578]: E1003 13:13:56.961212 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44e11f2ae23fddbb737f7b00fdf3e82e831289bd9ebbaf9260098bbe8900732b\": container with ID starting with 44e11f2ae23fddbb737f7b00fdf3e82e831289bd9ebbaf9260098bbe8900732b not found: ID does not exist" containerID="44e11f2ae23fddbb737f7b00fdf3e82e831289bd9ebbaf9260098bbe8900732b" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.961261 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44e11f2ae23fddbb737f7b00fdf3e82e831289bd9ebbaf9260098bbe8900732b"} err="failed to get container status \"44e11f2ae23fddbb737f7b00fdf3e82e831289bd9ebbaf9260098bbe8900732b\": rpc error: code = NotFound desc = could not find container \"44e11f2ae23fddbb737f7b00fdf3e82e831289bd9ebbaf9260098bbe8900732b\": container with ID starting with 44e11f2ae23fddbb737f7b00fdf3e82e831289bd9ebbaf9260098bbe8900732b not found: ID does not exist" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.967696 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.977702 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:13:56 crc kubenswrapper[4578]: E1003 13:13:56.978104 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec6cb50d-e8ea-450c-ba3c-e718e03c41a3" containerName="glance-log" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.978116 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec6cb50d-e8ea-450c-ba3c-e718e03c41a3" containerName="glance-log" Oct 03 13:13:56 crc kubenswrapper[4578]: E1003 13:13:56.978159 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec6cb50d-e8ea-450c-ba3c-e718e03c41a3" containerName="glance-httpd" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.978165 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec6cb50d-e8ea-450c-ba3c-e718e03c41a3" containerName="glance-httpd" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.978323 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec6cb50d-e8ea-450c-ba3c-e718e03c41a3" containerName="glance-log" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.978342 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec6cb50d-e8ea-450c-ba3c-e718e03c41a3" containerName="glance-httpd" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.979271 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.983960 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.984840 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 03 13:13:56 crc kubenswrapper[4578]: I1003 13:13:56.985293 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.095700 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d110456-18c3-4369-9ca7-0efcf6a02d8b-config-data\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.095773 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d110456-18c3-4369-9ca7-0efcf6a02d8b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.095797 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27ptw\" (UniqueName: \"kubernetes.io/projected/3d110456-18c3-4369-9ca7-0efcf6a02d8b-kube-api-access-27ptw\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.095849 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3d110456-18c3-4369-9ca7-0efcf6a02d8b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.095886 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d110456-18c3-4369-9ca7-0efcf6a02d8b-logs\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.095910 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d110456-18c3-4369-9ca7-0efcf6a02d8b-scripts\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.095966 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d110456-18c3-4369-9ca7-0efcf6a02d8b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.095996 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.198060 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3d110456-18c3-4369-9ca7-0efcf6a02d8b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.198114 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d110456-18c3-4369-9ca7-0efcf6a02d8b-logs\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.198145 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d110456-18c3-4369-9ca7-0efcf6a02d8b-scripts\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.198199 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d110456-18c3-4369-9ca7-0efcf6a02d8b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.198224 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.198268 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d110456-18c3-4369-9ca7-0efcf6a02d8b-config-data\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.198319 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d110456-18c3-4369-9ca7-0efcf6a02d8b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.198339 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27ptw\" (UniqueName: \"kubernetes.io/projected/3d110456-18c3-4369-9ca7-0efcf6a02d8b-kube-api-access-27ptw\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.198837 4578 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.198966 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d110456-18c3-4369-9ca7-0efcf6a02d8b-logs\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.199129 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3d110456-18c3-4369-9ca7-0efcf6a02d8b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.205348 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d110456-18c3-4369-9ca7-0efcf6a02d8b-scripts\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.205455 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d110456-18c3-4369-9ca7-0efcf6a02d8b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.206827 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d110456-18c3-4369-9ca7-0efcf6a02d8b-config-data\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.211920 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d110456-18c3-4369-9ca7-0efcf6a02d8b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.224762 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27ptw\" (UniqueName: \"kubernetes.io/projected/3d110456-18c3-4369-9ca7-0efcf6a02d8b-kube-api-access-27ptw\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.233460 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"3d110456-18c3-4369-9ca7-0efcf6a02d8b\") " pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.295885 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.683818 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.861706 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a925d503-688a-4a63-a6d5-1533a9a22aac","Type":"ContainerStarted","Data":"9c2664687a2e5ce8899e53ba9c5dfffce4a6572d01dd5e672b8a756e71a4722a"} Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.862075 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerName="ceilometer-central-agent" containerID="cri-o://8e9ae4feb2c38fbae277c3dbbdc74ea68b47414938aebd6eabba698f4914b621" gracePeriod=30 Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.862198 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerName="sg-core" containerID="cri-o://6853ac4b339f037dbd2c66955d65aaca7d497983ca6a0d31ff77dc5e57434c4e" gracePeriod=30 Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.862316 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerName="proxy-httpd" containerID="cri-o://9c2664687a2e5ce8899e53ba9c5dfffce4a6572d01dd5e672b8a756e71a4722a" gracePeriod=30 Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.862383 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerName="ceilometer-notification-agent" containerID="cri-o://6b5fa4fe8900e80130f7c569ae25f05762f0b391172a3db23a44840feff5ad7d" gracePeriod=30 Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.862530 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.899207 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.144613382 podStartE2EDuration="5.899184994s" podCreationTimestamp="2025-10-03 13:13:52 +0000 UTC" firstStartedPulling="2025-10-03 13:13:53.62728366 +0000 UTC m=+1369.425755844" lastFinishedPulling="2025-10-03 13:13:57.381855272 +0000 UTC m=+1373.180327456" observedRunningTime="2025-10-03 13:13:57.885972161 +0000 UTC m=+1373.684444335" watchObservedRunningTime="2025-10-03 13:13:57.899184994 +0000 UTC m=+1373.697657178" Oct 03 13:13:57 crc kubenswrapper[4578]: I1003 13:13:57.947558 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.386460 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sdwwm" Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.567630 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcwpf\" (UniqueName: \"kubernetes.io/projected/5533c61e-c654-4d79-9431-88bc37cff6d8-kube-api-access-lcwpf\") pod \"5533c61e-c654-4d79-9431-88bc37cff6d8\" (UID: \"5533c61e-c654-4d79-9431-88bc37cff6d8\") " Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.611080 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5533c61e-c654-4d79-9431-88bc37cff6d8-kube-api-access-lcwpf" (OuterVolumeSpecName: "kube-api-access-lcwpf") pod "5533c61e-c654-4d79-9431-88bc37cff6d8" (UID: "5533c61e-c654-4d79-9431-88bc37cff6d8"). InnerVolumeSpecName "kube-api-access-lcwpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.672936 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcwpf\" (UniqueName: \"kubernetes.io/projected/5533c61e-c654-4d79-9431-88bc37cff6d8-kube-api-access-lcwpf\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.748070 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4t996" Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.754920 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-56phj" Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.879289 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-459br\" (UniqueName: \"kubernetes.io/projected/cafa819d-515a-4e7a-8f97-e97522c4b31c-kube-api-access-459br\") pod \"cafa819d-515a-4e7a-8f97-e97522c4b31c\" (UID: \"cafa819d-515a-4e7a-8f97-e97522c4b31c\") " Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.879566 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58nkm\" (UniqueName: \"kubernetes.io/projected/9f94584d-3f52-4dcc-9c36-7f1174df9753-kube-api-access-58nkm\") pod \"9f94584d-3f52-4dcc-9c36-7f1174df9753\" (UID: \"9f94584d-3f52-4dcc-9c36-7f1174df9753\") " Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.885697 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cafa819d-515a-4e7a-8f97-e97522c4b31c-kube-api-access-459br" (OuterVolumeSpecName: "kube-api-access-459br") pod "cafa819d-515a-4e7a-8f97-e97522c4b31c" (UID: "cafa819d-515a-4e7a-8f97-e97522c4b31c"). InnerVolumeSpecName "kube-api-access-459br". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.891672 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4t996" Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.892872 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4t996" event={"ID":"9f94584d-3f52-4dcc-9c36-7f1174df9753","Type":"ContainerDied","Data":"45eaf9479201e417fa9605615adcbaaf3f43673c6dc34090273e7df06bc89649"} Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.892900 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45eaf9479201e417fa9605615adcbaaf3f43673c6dc34090273e7df06bc89649" Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.894054 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f94584d-3f52-4dcc-9c36-7f1174df9753-kube-api-access-58nkm" (OuterVolumeSpecName: "kube-api-access-58nkm") pod "9f94584d-3f52-4dcc-9c36-7f1174df9753" (UID: "9f94584d-3f52-4dcc-9c36-7f1174df9753"). InnerVolumeSpecName "kube-api-access-58nkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.904076 4578 generic.go:334] "Generic (PLEG): container finished" podID="25ef994a-0092-4e72-a636-f51bbd1a8a99" containerID="2323557f12cf7f4bfe4a74ca5af7e4ce03b61bb618e398e5eba8cfa1c269bcda" exitCode=0 Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.904170 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"25ef994a-0092-4e72-a636-f51bbd1a8a99","Type":"ContainerDied","Data":"2323557f12cf7f4bfe4a74ca5af7e4ce03b61bb618e398e5eba8cfa1c269bcda"} Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.924904 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sdwwm" Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.938295 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec6cb50d-e8ea-450c-ba3c-e718e03c41a3" path="/var/lib/kubelet/pods/ec6cb50d-e8ea-450c-ba3c-e718e03c41a3/volumes" Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.939366 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-sdwwm" event={"ID":"5533c61e-c654-4d79-9431-88bc37cff6d8","Type":"ContainerDied","Data":"7475f482cf997546210082ce0180d9e0ab5ba20c5af3520d84f2331a16c340fa"} Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.939390 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7475f482cf997546210082ce0180d9e0ab5ba20c5af3520d84f2331a16c340fa" Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.949723 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-56phj" Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.949708 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-56phj" event={"ID":"cafa819d-515a-4e7a-8f97-e97522c4b31c","Type":"ContainerDied","Data":"fa9ebff312d6df85772f1d9c2b2574895c5445f2785be9dbe0207f1eb6a06923"} Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.949831 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa9ebff312d6df85772f1d9c2b2574895c5445f2785be9dbe0207f1eb6a06923" Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.953540 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3d110456-18c3-4369-9ca7-0efcf6a02d8b","Type":"ContainerStarted","Data":"f70192e890d0fd13308e9ca633200e45caadc2ca0430964c5d51de9045566732"} Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.968183 4578 generic.go:334] "Generic (PLEG): container finished" podID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerID="6853ac4b339f037dbd2c66955d65aaca7d497983ca6a0d31ff77dc5e57434c4e" exitCode=2 Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.968217 4578 generic.go:334] "Generic (PLEG): container finished" podID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerID="6b5fa4fe8900e80130f7c569ae25f05762f0b391172a3db23a44840feff5ad7d" exitCode=0 Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.968237 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a925d503-688a-4a63-a6d5-1533a9a22aac","Type":"ContainerDied","Data":"6853ac4b339f037dbd2c66955d65aaca7d497983ca6a0d31ff77dc5e57434c4e"} Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.968261 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a925d503-688a-4a63-a6d5-1533a9a22aac","Type":"ContainerDied","Data":"6b5fa4fe8900e80130f7c569ae25f05762f0b391172a3db23a44840feff5ad7d"} Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.982000 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58nkm\" (UniqueName: \"kubernetes.io/projected/9f94584d-3f52-4dcc-9c36-7f1174df9753-kube-api-access-58nkm\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:58 crc kubenswrapper[4578]: I1003 13:13:58.982030 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-459br\" (UniqueName: \"kubernetes.io/projected/cafa819d-515a-4e7a-8f97-e97522c4b31c-kube-api-access-459br\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.098624 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.286422 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-combined-ca-bundle\") pod \"25ef994a-0092-4e72-a636-f51bbd1a8a99\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.286477 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-config-data\") pod \"25ef994a-0092-4e72-a636-f51bbd1a8a99\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.286542 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-scripts\") pod \"25ef994a-0092-4e72-a636-f51bbd1a8a99\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.286585 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gjc7\" (UniqueName: \"kubernetes.io/projected/25ef994a-0092-4e72-a636-f51bbd1a8a99-kube-api-access-5gjc7\") pod \"25ef994a-0092-4e72-a636-f51bbd1a8a99\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.286682 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/25ef994a-0092-4e72-a636-f51bbd1a8a99-httpd-run\") pod \"25ef994a-0092-4e72-a636-f51bbd1a8a99\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.286741 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25ef994a-0092-4e72-a636-f51bbd1a8a99-logs\") pod \"25ef994a-0092-4e72-a636-f51bbd1a8a99\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.286763 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"25ef994a-0092-4e72-a636-f51bbd1a8a99\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.286789 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-internal-tls-certs\") pod \"25ef994a-0092-4e72-a636-f51bbd1a8a99\" (UID: \"25ef994a-0092-4e72-a636-f51bbd1a8a99\") " Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.293853 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25ef994a-0092-4e72-a636-f51bbd1a8a99-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "25ef994a-0092-4e72-a636-f51bbd1a8a99" (UID: "25ef994a-0092-4e72-a636-f51bbd1a8a99"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.294169 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25ef994a-0092-4e72-a636-f51bbd1a8a99-logs" (OuterVolumeSpecName: "logs") pod "25ef994a-0092-4e72-a636-f51bbd1a8a99" (UID: "25ef994a-0092-4e72-a636-f51bbd1a8a99"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.295741 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-scripts" (OuterVolumeSpecName: "scripts") pod "25ef994a-0092-4e72-a636-f51bbd1a8a99" (UID: "25ef994a-0092-4e72-a636-f51bbd1a8a99"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.296019 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25ef994a-0092-4e72-a636-f51bbd1a8a99-kube-api-access-5gjc7" (OuterVolumeSpecName: "kube-api-access-5gjc7") pod "25ef994a-0092-4e72-a636-f51bbd1a8a99" (UID: "25ef994a-0092-4e72-a636-f51bbd1a8a99"). InnerVolumeSpecName "kube-api-access-5gjc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.299945 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "25ef994a-0092-4e72-a636-f51bbd1a8a99" (UID: "25ef994a-0092-4e72-a636-f51bbd1a8a99"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.320440 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25ef994a-0092-4e72-a636-f51bbd1a8a99" (UID: "25ef994a-0092-4e72-a636-f51bbd1a8a99"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.357552 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "25ef994a-0092-4e72-a636-f51bbd1a8a99" (UID: "25ef994a-0092-4e72-a636-f51bbd1a8a99"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.370932 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-config-data" (OuterVolumeSpecName: "config-data") pod "25ef994a-0092-4e72-a636-f51bbd1a8a99" (UID: "25ef994a-0092-4e72-a636-f51bbd1a8a99"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.389584 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gjc7\" (UniqueName: \"kubernetes.io/projected/25ef994a-0092-4e72-a636-f51bbd1a8a99-kube-api-access-5gjc7\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.389624 4578 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/25ef994a-0092-4e72-a636-f51bbd1a8a99-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.389651 4578 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25ef994a-0092-4e72-a636-f51bbd1a8a99-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.389685 4578 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.389697 4578 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.389705 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.389713 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.389721 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25ef994a-0092-4e72-a636-f51bbd1a8a99-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.411194 4578 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.491102 4578 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.978616 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.978614 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"25ef994a-0092-4e72-a636-f51bbd1a8a99","Type":"ContainerDied","Data":"cc772551e8d3098cae7f8cc16d31020fbd3c6af7e2c008694afe493d7a5e0cc1"} Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.978784 4578 scope.go:117] "RemoveContainer" containerID="2323557f12cf7f4bfe4a74ca5af7e4ce03b61bb618e398e5eba8cfa1c269bcda" Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.984594 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3d110456-18c3-4369-9ca7-0efcf6a02d8b","Type":"ContainerStarted","Data":"37d9b507f06d29f3d090e4aa3148421194f00f51ce63c8e41732988944f4b444"} Oct 03 13:13:59 crc kubenswrapper[4578]: I1003 13:13:59.984646 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3d110456-18c3-4369-9ca7-0efcf6a02d8b","Type":"ContainerStarted","Data":"c5aa5c3a6fe6b92ccb4d33c0e1a6dae7fd91f447708519156d8f543f390d987f"} Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.015184 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.015168762 podStartE2EDuration="4.015168762s" podCreationTimestamp="2025-10-03 13:13:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:14:00.013771628 +0000 UTC m=+1375.812243812" watchObservedRunningTime="2025-10-03 13:14:00.015168762 +0000 UTC m=+1375.813640946" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.029810 4578 scope.go:117] "RemoveContainer" containerID="666b5fd3073ef524529191139ec8cb50cfbb3d1d8b1a4b19cea1ea993e443544" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.054201 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.067521 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.074719 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:14:00 crc kubenswrapper[4578]: E1003 13:14:00.075069 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25ef994a-0092-4e72-a636-f51bbd1a8a99" containerName="glance-httpd" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.075086 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="25ef994a-0092-4e72-a636-f51bbd1a8a99" containerName="glance-httpd" Oct 03 13:14:00 crc kubenswrapper[4578]: E1003 13:14:00.075105 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5533c61e-c654-4d79-9431-88bc37cff6d8" containerName="mariadb-database-create" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.075111 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="5533c61e-c654-4d79-9431-88bc37cff6d8" containerName="mariadb-database-create" Oct 03 13:14:00 crc kubenswrapper[4578]: E1003 13:14:00.075122 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cafa819d-515a-4e7a-8f97-e97522c4b31c" containerName="mariadb-database-create" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.075129 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="cafa819d-515a-4e7a-8f97-e97522c4b31c" containerName="mariadb-database-create" Oct 03 13:14:00 crc kubenswrapper[4578]: E1003 13:14:00.075137 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25ef994a-0092-4e72-a636-f51bbd1a8a99" containerName="glance-log" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.075143 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="25ef994a-0092-4e72-a636-f51bbd1a8a99" containerName="glance-log" Oct 03 13:14:00 crc kubenswrapper[4578]: E1003 13:14:00.075157 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f94584d-3f52-4dcc-9c36-7f1174df9753" containerName="mariadb-database-create" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.075164 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f94584d-3f52-4dcc-9c36-7f1174df9753" containerName="mariadb-database-create" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.075428 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="cafa819d-515a-4e7a-8f97-e97522c4b31c" containerName="mariadb-database-create" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.075442 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="25ef994a-0092-4e72-a636-f51bbd1a8a99" containerName="glance-log" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.075451 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="5533c61e-c654-4d79-9431-88bc37cff6d8" containerName="mariadb-database-create" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.075466 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f94584d-3f52-4dcc-9c36-7f1174df9753" containerName="mariadb-database-create" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.075476 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="25ef994a-0092-4e72-a636-f51bbd1a8a99" containerName="glance-httpd" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.076913 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.080687 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.085721 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.103064 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.203092 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv8gr\" (UniqueName: \"kubernetes.io/projected/92636290-70b5-44e9-94f2-988b490ae46c-kube-api-access-fv8gr\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.203132 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92636290-70b5-44e9-94f2-988b490ae46c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.203202 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/92636290-70b5-44e9-94f2-988b490ae46c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.203268 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92636290-70b5-44e9-94f2-988b490ae46c-logs\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.203283 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.203305 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92636290-70b5-44e9-94f2-988b490ae46c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.203333 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/92636290-70b5-44e9-94f2-988b490ae46c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.203375 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92636290-70b5-44e9-94f2-988b490ae46c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.304801 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/92636290-70b5-44e9-94f2-988b490ae46c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.304871 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92636290-70b5-44e9-94f2-988b490ae46c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.304919 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv8gr\" (UniqueName: \"kubernetes.io/projected/92636290-70b5-44e9-94f2-988b490ae46c-kube-api-access-fv8gr\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.304939 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92636290-70b5-44e9-94f2-988b490ae46c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.304987 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/92636290-70b5-44e9-94f2-988b490ae46c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.305039 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92636290-70b5-44e9-94f2-988b490ae46c-logs\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.305058 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.305078 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92636290-70b5-44e9-94f2-988b490ae46c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.306027 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92636290-70b5-44e9-94f2-988b490ae46c-logs\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.306283 4578 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.306428 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/92636290-70b5-44e9-94f2-988b490ae46c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.314843 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92636290-70b5-44e9-94f2-988b490ae46c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.322493 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92636290-70b5-44e9-94f2-988b490ae46c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.324316 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/92636290-70b5-44e9-94f2-988b490ae46c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.324970 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92636290-70b5-44e9-94f2-988b490ae46c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.334763 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv8gr\" (UniqueName: \"kubernetes.io/projected/92636290-70b5-44e9-94f2-988b490ae46c-kube-api-access-fv8gr\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.371459 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"92636290-70b5-44e9-94f2-988b490ae46c\") " pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.401190 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 03 13:14:00 crc kubenswrapper[4578]: I1003 13:14:00.922718 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25ef994a-0092-4e72-a636-f51bbd1a8a99" path="/var/lib/kubelet/pods/25ef994a-0092-4e72-a636-f51bbd1a8a99/volumes" Oct 03 13:14:01 crc kubenswrapper[4578]: I1003 13:14:01.021130 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"92636290-70b5-44e9-94f2-988b490ae46c","Type":"ContainerStarted","Data":"c72baae902f01f2b5c74a54d50bad5babcbd54e38198543f36f1519ed90973ec"} Oct 03 13:14:01 crc kubenswrapper[4578]: I1003 13:14:01.022798 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 03 13:14:02 crc kubenswrapper[4578]: I1003 13:14:02.055145 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"92636290-70b5-44e9-94f2-988b490ae46c","Type":"ContainerStarted","Data":"37f53dfb0d010b79efb5045f8fe7860d06c9ecc2c90b69e993079cd48cd50c6d"} Oct 03 13:14:03 crc kubenswrapper[4578]: I1003 13:14:03.064831 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"92636290-70b5-44e9-94f2-988b490ae46c","Type":"ContainerStarted","Data":"6dae7c391059eedd88fc216451d17a017bd24a851f1b3de4b03f22ac8c91b016"} Oct 03 13:14:03 crc kubenswrapper[4578]: I1003 13:14:03.085518 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.085501811 podStartE2EDuration="3.085501811s" podCreationTimestamp="2025-10-03 13:14:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:14:03.081099943 +0000 UTC m=+1378.879572137" watchObservedRunningTime="2025-10-03 13:14:03.085501811 +0000 UTC m=+1378.883973995" Oct 03 13:14:04 crc kubenswrapper[4578]: I1003 13:14:04.077188 4578 generic.go:334] "Generic (PLEG): container finished" podID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerID="8e9ae4feb2c38fbae277c3dbbdc74ea68b47414938aebd6eabba698f4914b621" exitCode=0 Oct 03 13:14:04 crc kubenswrapper[4578]: I1003 13:14:04.077257 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a925d503-688a-4a63-a6d5-1533a9a22aac","Type":"ContainerDied","Data":"8e9ae4feb2c38fbae277c3dbbdc74ea68b47414938aebd6eabba698f4914b621"} Oct 03 13:14:04 crc kubenswrapper[4578]: I1003 13:14:04.848824 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-9d21-account-create-47c2p"] Oct 03 13:14:04 crc kubenswrapper[4578]: I1003 13:14:04.850029 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9d21-account-create-47c2p" Oct 03 13:14:04 crc kubenswrapper[4578]: I1003 13:14:04.853199 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 03 13:14:04 crc kubenswrapper[4578]: I1003 13:14:04.869710 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-9d21-account-create-47c2p"] Oct 03 13:14:04 crc kubenswrapper[4578]: I1003 13:14:04.918227 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtzdl\" (UniqueName: \"kubernetes.io/projected/c9cd3675-bb86-4dbd-8da1-a25d0fe30298-kube-api-access-vtzdl\") pod \"nova-api-9d21-account-create-47c2p\" (UID: \"c9cd3675-bb86-4dbd-8da1-a25d0fe30298\") " pod="openstack/nova-api-9d21-account-create-47c2p" Oct 03 13:14:05 crc kubenswrapper[4578]: I1003 13:14:05.020748 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtzdl\" (UniqueName: \"kubernetes.io/projected/c9cd3675-bb86-4dbd-8da1-a25d0fe30298-kube-api-access-vtzdl\") pod \"nova-api-9d21-account-create-47c2p\" (UID: \"c9cd3675-bb86-4dbd-8da1-a25d0fe30298\") " pod="openstack/nova-api-9d21-account-create-47c2p" Oct 03 13:14:05 crc kubenswrapper[4578]: I1003 13:14:05.053302 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtzdl\" (UniqueName: \"kubernetes.io/projected/c9cd3675-bb86-4dbd-8da1-a25d0fe30298-kube-api-access-vtzdl\") pod \"nova-api-9d21-account-create-47c2p\" (UID: \"c9cd3675-bb86-4dbd-8da1-a25d0fe30298\") " pod="openstack/nova-api-9d21-account-create-47c2p" Oct 03 13:14:05 crc kubenswrapper[4578]: I1003 13:14:05.063995 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-5bf9-account-create-9vgfl"] Oct 03 13:14:05 crc kubenswrapper[4578]: I1003 13:14:05.066906 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5bf9-account-create-9vgfl" Oct 03 13:14:05 crc kubenswrapper[4578]: I1003 13:14:05.073243 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 03 13:14:05 crc kubenswrapper[4578]: I1003 13:14:05.085200 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-5bf9-account-create-9vgfl"] Oct 03 13:14:05 crc kubenswrapper[4578]: I1003 13:14:05.122401 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnvkn\" (UniqueName: \"kubernetes.io/projected/a4229058-7b31-4a35-8ddc-6db4cf3da084-kube-api-access-hnvkn\") pod \"nova-cell0-5bf9-account-create-9vgfl\" (UID: \"a4229058-7b31-4a35-8ddc-6db4cf3da084\") " pod="openstack/nova-cell0-5bf9-account-create-9vgfl" Oct 03 13:14:05 crc kubenswrapper[4578]: I1003 13:14:05.151106 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-b3d4-account-create-nxpl4"] Oct 03 13:14:05 crc kubenswrapper[4578]: I1003 13:14:05.152241 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b3d4-account-create-nxpl4" Oct 03 13:14:05 crc kubenswrapper[4578]: I1003 13:14:05.156103 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 03 13:14:05 crc kubenswrapper[4578]: I1003 13:14:05.163951 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-b3d4-account-create-nxpl4"] Oct 03 13:14:05 crc kubenswrapper[4578]: I1003 13:14:05.170848 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9d21-account-create-47c2p" Oct 03 13:14:05 crc kubenswrapper[4578]: I1003 13:14:05.226364 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s69hw\" (UniqueName: \"kubernetes.io/projected/166d9486-f455-4541-b53c-89b7203f6994-kube-api-access-s69hw\") pod \"nova-cell1-b3d4-account-create-nxpl4\" (UID: \"166d9486-f455-4541-b53c-89b7203f6994\") " pod="openstack/nova-cell1-b3d4-account-create-nxpl4" Oct 03 13:14:05 crc kubenswrapper[4578]: I1003 13:14:05.226453 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnvkn\" (UniqueName: \"kubernetes.io/projected/a4229058-7b31-4a35-8ddc-6db4cf3da084-kube-api-access-hnvkn\") pod \"nova-cell0-5bf9-account-create-9vgfl\" (UID: \"a4229058-7b31-4a35-8ddc-6db4cf3da084\") " pod="openstack/nova-cell0-5bf9-account-create-9vgfl" Oct 03 13:14:05 crc kubenswrapper[4578]: I1003 13:14:05.249179 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnvkn\" (UniqueName: \"kubernetes.io/projected/a4229058-7b31-4a35-8ddc-6db4cf3da084-kube-api-access-hnvkn\") pod \"nova-cell0-5bf9-account-create-9vgfl\" (UID: \"a4229058-7b31-4a35-8ddc-6db4cf3da084\") " pod="openstack/nova-cell0-5bf9-account-create-9vgfl" Oct 03 13:14:05 crc kubenswrapper[4578]: I1003 13:14:05.328871 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s69hw\" (UniqueName: \"kubernetes.io/projected/166d9486-f455-4541-b53c-89b7203f6994-kube-api-access-s69hw\") pod \"nova-cell1-b3d4-account-create-nxpl4\" (UID: \"166d9486-f455-4541-b53c-89b7203f6994\") " pod="openstack/nova-cell1-b3d4-account-create-nxpl4" Oct 03 13:14:05 crc kubenswrapper[4578]: I1003 13:14:05.360321 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s69hw\" (UniqueName: \"kubernetes.io/projected/166d9486-f455-4541-b53c-89b7203f6994-kube-api-access-s69hw\") pod \"nova-cell1-b3d4-account-create-nxpl4\" (UID: \"166d9486-f455-4541-b53c-89b7203f6994\") " pod="openstack/nova-cell1-b3d4-account-create-nxpl4" Oct 03 13:14:05 crc kubenswrapper[4578]: I1003 13:14:05.433764 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5bf9-account-create-9vgfl" Oct 03 13:14:05 crc kubenswrapper[4578]: I1003 13:14:05.477107 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b3d4-account-create-nxpl4" Oct 03 13:14:05 crc kubenswrapper[4578]: I1003 13:14:05.757204 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-9d21-account-create-47c2p"] Oct 03 13:14:05 crc kubenswrapper[4578]: W1003 13:14:05.769605 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9cd3675_bb86_4dbd_8da1_a25d0fe30298.slice/crio-64036f3252cfc4d7ecb001be5c8e7eeedbdca1ea46ced0ec6c59aaa45418cfcd WatchSource:0}: Error finding container 64036f3252cfc4d7ecb001be5c8e7eeedbdca1ea46ced0ec6c59aaa45418cfcd: Status 404 returned error can't find the container with id 64036f3252cfc4d7ecb001be5c8e7eeedbdca1ea46ced0ec6c59aaa45418cfcd Oct 03 13:14:06 crc kubenswrapper[4578]: I1003 13:14:06.025383 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-b3d4-account-create-nxpl4"] Oct 03 13:14:06 crc kubenswrapper[4578]: I1003 13:14:06.040483 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-5bf9-account-create-9vgfl"] Oct 03 13:14:06 crc kubenswrapper[4578]: W1003 13:14:06.042849 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4229058_7b31_4a35_8ddc_6db4cf3da084.slice/crio-d179b627f3d6252526f4cca0ebc8ea954a68eb7cbca6f30077855058b0e9b031 WatchSource:0}: Error finding container d179b627f3d6252526f4cca0ebc8ea954a68eb7cbca6f30077855058b0e9b031: Status 404 returned error can't find the container with id d179b627f3d6252526f4cca0ebc8ea954a68eb7cbca6f30077855058b0e9b031 Oct 03 13:14:06 crc kubenswrapper[4578]: I1003 13:14:06.130778 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5bf9-account-create-9vgfl" event={"ID":"a4229058-7b31-4a35-8ddc-6db4cf3da084","Type":"ContainerStarted","Data":"d179b627f3d6252526f4cca0ebc8ea954a68eb7cbca6f30077855058b0e9b031"} Oct 03 13:14:06 crc kubenswrapper[4578]: I1003 13:14:06.135600 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b3d4-account-create-nxpl4" event={"ID":"166d9486-f455-4541-b53c-89b7203f6994","Type":"ContainerStarted","Data":"70d7c2df405825fa5e19229986fed29320e94bdc736c3e9bcaca4c276aa69648"} Oct 03 13:14:06 crc kubenswrapper[4578]: I1003 13:14:06.137802 4578 generic.go:334] "Generic (PLEG): container finished" podID="c9cd3675-bb86-4dbd-8da1-a25d0fe30298" containerID="eba716b6694fb5f7bc63e249972dd056be854b91df8c6c802660a17d59055da2" exitCode=0 Oct 03 13:14:06 crc kubenswrapper[4578]: I1003 13:14:06.137849 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-9d21-account-create-47c2p" event={"ID":"c9cd3675-bb86-4dbd-8da1-a25d0fe30298","Type":"ContainerDied","Data":"eba716b6694fb5f7bc63e249972dd056be854b91df8c6c802660a17d59055da2"} Oct 03 13:14:06 crc kubenswrapper[4578]: I1003 13:14:06.137874 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-9d21-account-create-47c2p" event={"ID":"c9cd3675-bb86-4dbd-8da1-a25d0fe30298","Type":"ContainerStarted","Data":"64036f3252cfc4d7ecb001be5c8e7eeedbdca1ea46ced0ec6c59aaa45418cfcd"} Oct 03 13:14:06 crc kubenswrapper[4578]: I1003 13:14:06.781729 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7b58cf9568-f8nng" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 03 13:14:06 crc kubenswrapper[4578]: I1003 13:14:06.948562 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-68684899bd-2cwll" podUID="d49d623e-2cd6-4786-b1da-34b78b89a134" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 03 13:14:07 crc kubenswrapper[4578]: I1003 13:14:07.170680 4578 generic.go:334] "Generic (PLEG): container finished" podID="a4229058-7b31-4a35-8ddc-6db4cf3da084" containerID="242a9c8a497def7c5a681fa38a2aedd4beda4d7b738bbbbb43de23ec150a1fcb" exitCode=0 Oct 03 13:14:07 crc kubenswrapper[4578]: I1003 13:14:07.170769 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5bf9-account-create-9vgfl" event={"ID":"a4229058-7b31-4a35-8ddc-6db4cf3da084","Type":"ContainerDied","Data":"242a9c8a497def7c5a681fa38a2aedd4beda4d7b738bbbbb43de23ec150a1fcb"} Oct 03 13:14:07 crc kubenswrapper[4578]: I1003 13:14:07.173387 4578 generic.go:334] "Generic (PLEG): container finished" podID="166d9486-f455-4541-b53c-89b7203f6994" containerID="359723667c8d4494fbfb2c97e0d75ca723efa35033aff65f5acddc2efdfab4f6" exitCode=0 Oct 03 13:14:07 crc kubenswrapper[4578]: I1003 13:14:07.174043 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b3d4-account-create-nxpl4" event={"ID":"166d9486-f455-4541-b53c-89b7203f6994","Type":"ContainerDied","Data":"359723667c8d4494fbfb2c97e0d75ca723efa35033aff65f5acddc2efdfab4f6"} Oct 03 13:14:07 crc kubenswrapper[4578]: I1003 13:14:07.296853 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 03 13:14:07 crc kubenswrapper[4578]: I1003 13:14:07.297321 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 03 13:14:07 crc kubenswrapper[4578]: I1003 13:14:07.346497 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 03 13:14:07 crc kubenswrapper[4578]: I1003 13:14:07.359254 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 03 13:14:07 crc kubenswrapper[4578]: I1003 13:14:07.637778 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9d21-account-create-47c2p" Oct 03 13:14:07 crc kubenswrapper[4578]: I1003 13:14:07.683018 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtzdl\" (UniqueName: \"kubernetes.io/projected/c9cd3675-bb86-4dbd-8da1-a25d0fe30298-kube-api-access-vtzdl\") pod \"c9cd3675-bb86-4dbd-8da1-a25d0fe30298\" (UID: \"c9cd3675-bb86-4dbd-8da1-a25d0fe30298\") " Oct 03 13:14:07 crc kubenswrapper[4578]: I1003 13:14:07.695298 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9cd3675-bb86-4dbd-8da1-a25d0fe30298-kube-api-access-vtzdl" (OuterVolumeSpecName: "kube-api-access-vtzdl") pod "c9cd3675-bb86-4dbd-8da1-a25d0fe30298" (UID: "c9cd3675-bb86-4dbd-8da1-a25d0fe30298"). InnerVolumeSpecName "kube-api-access-vtzdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:14:07 crc kubenswrapper[4578]: I1003 13:14:07.785191 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtzdl\" (UniqueName: \"kubernetes.io/projected/c9cd3675-bb86-4dbd-8da1-a25d0fe30298-kube-api-access-vtzdl\") on node \"crc\" DevicePath \"\"" Oct 03 13:14:08 crc kubenswrapper[4578]: I1003 13:14:08.184104 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-9d21-account-create-47c2p" event={"ID":"c9cd3675-bb86-4dbd-8da1-a25d0fe30298","Type":"ContainerDied","Data":"64036f3252cfc4d7ecb001be5c8e7eeedbdca1ea46ced0ec6c59aaa45418cfcd"} Oct 03 13:14:08 crc kubenswrapper[4578]: I1003 13:14:08.184143 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64036f3252cfc4d7ecb001be5c8e7eeedbdca1ea46ced0ec6c59aaa45418cfcd" Oct 03 13:14:08 crc kubenswrapper[4578]: I1003 13:14:08.184204 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9d21-account-create-47c2p" Oct 03 13:14:08 crc kubenswrapper[4578]: I1003 13:14:08.185021 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 03 13:14:08 crc kubenswrapper[4578]: I1003 13:14:08.185308 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 03 13:14:08 crc kubenswrapper[4578]: I1003 13:14:08.732470 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5bf9-account-create-9vgfl" Oct 03 13:14:08 crc kubenswrapper[4578]: I1003 13:14:08.802736 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnvkn\" (UniqueName: \"kubernetes.io/projected/a4229058-7b31-4a35-8ddc-6db4cf3da084-kube-api-access-hnvkn\") pod \"a4229058-7b31-4a35-8ddc-6db4cf3da084\" (UID: \"a4229058-7b31-4a35-8ddc-6db4cf3da084\") " Oct 03 13:14:08 crc kubenswrapper[4578]: I1003 13:14:08.808990 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4229058-7b31-4a35-8ddc-6db4cf3da084-kube-api-access-hnvkn" (OuterVolumeSpecName: "kube-api-access-hnvkn") pod "a4229058-7b31-4a35-8ddc-6db4cf3da084" (UID: "a4229058-7b31-4a35-8ddc-6db4cf3da084"). InnerVolumeSpecName "kube-api-access-hnvkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:14:08 crc kubenswrapper[4578]: I1003 13:14:08.875511 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b3d4-account-create-nxpl4" Oct 03 13:14:08 crc kubenswrapper[4578]: I1003 13:14:08.904530 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s69hw\" (UniqueName: \"kubernetes.io/projected/166d9486-f455-4541-b53c-89b7203f6994-kube-api-access-s69hw\") pod \"166d9486-f455-4541-b53c-89b7203f6994\" (UID: \"166d9486-f455-4541-b53c-89b7203f6994\") " Oct 03 13:14:08 crc kubenswrapper[4578]: I1003 13:14:08.905040 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnvkn\" (UniqueName: \"kubernetes.io/projected/a4229058-7b31-4a35-8ddc-6db4cf3da084-kube-api-access-hnvkn\") on node \"crc\" DevicePath \"\"" Oct 03 13:14:08 crc kubenswrapper[4578]: I1003 13:14:08.909073 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/166d9486-f455-4541-b53c-89b7203f6994-kube-api-access-s69hw" (OuterVolumeSpecName: "kube-api-access-s69hw") pod "166d9486-f455-4541-b53c-89b7203f6994" (UID: "166d9486-f455-4541-b53c-89b7203f6994"). InnerVolumeSpecName "kube-api-access-s69hw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:14:09 crc kubenswrapper[4578]: I1003 13:14:09.006220 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s69hw\" (UniqueName: \"kubernetes.io/projected/166d9486-f455-4541-b53c-89b7203f6994-kube-api-access-s69hw\") on node \"crc\" DevicePath \"\"" Oct 03 13:14:09 crc kubenswrapper[4578]: I1003 13:14:09.196373 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5bf9-account-create-9vgfl" event={"ID":"a4229058-7b31-4a35-8ddc-6db4cf3da084","Type":"ContainerDied","Data":"d179b627f3d6252526f4cca0ebc8ea954a68eb7cbca6f30077855058b0e9b031"} Oct 03 13:14:09 crc kubenswrapper[4578]: I1003 13:14:09.196423 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5bf9-account-create-9vgfl" Oct 03 13:14:09 crc kubenswrapper[4578]: I1003 13:14:09.196431 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d179b627f3d6252526f4cca0ebc8ea954a68eb7cbca6f30077855058b0e9b031" Oct 03 13:14:09 crc kubenswrapper[4578]: I1003 13:14:09.202587 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b3d4-account-create-nxpl4" event={"ID":"166d9486-f455-4541-b53c-89b7203f6994","Type":"ContainerDied","Data":"70d7c2df405825fa5e19229986fed29320e94bdc736c3e9bcaca4c276aa69648"} Oct 03 13:14:09 crc kubenswrapper[4578]: I1003 13:14:09.202659 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70d7c2df405825fa5e19229986fed29320e94bdc736c3e9bcaca4c276aa69648" Oct 03 13:14:09 crc kubenswrapper[4578]: I1003 13:14:09.202660 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b3d4-account-create-nxpl4" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.402437 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.403425 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.507127 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.615476 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.640212 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fmdjc"] Oct 03 13:14:10 crc kubenswrapper[4578]: E1003 13:14:10.640670 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4229058-7b31-4a35-8ddc-6db4cf3da084" containerName="mariadb-account-create" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.640691 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4229058-7b31-4a35-8ddc-6db4cf3da084" containerName="mariadb-account-create" Oct 03 13:14:10 crc kubenswrapper[4578]: E1003 13:14:10.640707 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9cd3675-bb86-4dbd-8da1-a25d0fe30298" containerName="mariadb-account-create" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.640716 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9cd3675-bb86-4dbd-8da1-a25d0fe30298" containerName="mariadb-account-create" Oct 03 13:14:10 crc kubenswrapper[4578]: E1003 13:14:10.640742 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="166d9486-f455-4541-b53c-89b7203f6994" containerName="mariadb-account-create" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.640749 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="166d9486-f455-4541-b53c-89b7203f6994" containerName="mariadb-account-create" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.640954 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="166d9486-f455-4541-b53c-89b7203f6994" containerName="mariadb-account-create" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.641001 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9cd3675-bb86-4dbd-8da1-a25d0fe30298" containerName="mariadb-account-create" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.641016 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4229058-7b31-4a35-8ddc-6db4cf3da084" containerName="mariadb-account-create" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.641723 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fmdjc" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.647324 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-8dm6s" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.647552 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.650935 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.657021 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fmdjc"] Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.747193 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f6e7467-e546-4563-bbe5-30d181f18349-scripts\") pod \"nova-cell0-conductor-db-sync-fmdjc\" (UID: \"5f6e7467-e546-4563-bbe5-30d181f18349\") " pod="openstack/nova-cell0-conductor-db-sync-fmdjc" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.747283 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f6e7467-e546-4563-bbe5-30d181f18349-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fmdjc\" (UID: \"5f6e7467-e546-4563-bbe5-30d181f18349\") " pod="openstack/nova-cell0-conductor-db-sync-fmdjc" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.747326 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f6e7467-e546-4563-bbe5-30d181f18349-config-data\") pod \"nova-cell0-conductor-db-sync-fmdjc\" (UID: \"5f6e7467-e546-4563-bbe5-30d181f18349\") " pod="openstack/nova-cell0-conductor-db-sync-fmdjc" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.747371 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gchdq\" (UniqueName: \"kubernetes.io/projected/5f6e7467-e546-4563-bbe5-30d181f18349-kube-api-access-gchdq\") pod \"nova-cell0-conductor-db-sync-fmdjc\" (UID: \"5f6e7467-e546-4563-bbe5-30d181f18349\") " pod="openstack/nova-cell0-conductor-db-sync-fmdjc" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.763315 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.763408 4578 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.835581 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.849925 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f6e7467-e546-4563-bbe5-30d181f18349-scripts\") pod \"nova-cell0-conductor-db-sync-fmdjc\" (UID: \"5f6e7467-e546-4563-bbe5-30d181f18349\") " pod="openstack/nova-cell0-conductor-db-sync-fmdjc" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.849999 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f6e7467-e546-4563-bbe5-30d181f18349-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fmdjc\" (UID: \"5f6e7467-e546-4563-bbe5-30d181f18349\") " pod="openstack/nova-cell0-conductor-db-sync-fmdjc" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.850048 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f6e7467-e546-4563-bbe5-30d181f18349-config-data\") pod \"nova-cell0-conductor-db-sync-fmdjc\" (UID: \"5f6e7467-e546-4563-bbe5-30d181f18349\") " pod="openstack/nova-cell0-conductor-db-sync-fmdjc" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.850081 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gchdq\" (UniqueName: \"kubernetes.io/projected/5f6e7467-e546-4563-bbe5-30d181f18349-kube-api-access-gchdq\") pod \"nova-cell0-conductor-db-sync-fmdjc\" (UID: \"5f6e7467-e546-4563-bbe5-30d181f18349\") " pod="openstack/nova-cell0-conductor-db-sync-fmdjc" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.858883 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f6e7467-e546-4563-bbe5-30d181f18349-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fmdjc\" (UID: \"5f6e7467-e546-4563-bbe5-30d181f18349\") " pod="openstack/nova-cell0-conductor-db-sync-fmdjc" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.863998 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f6e7467-e546-4563-bbe5-30d181f18349-scripts\") pod \"nova-cell0-conductor-db-sync-fmdjc\" (UID: \"5f6e7467-e546-4563-bbe5-30d181f18349\") " pod="openstack/nova-cell0-conductor-db-sync-fmdjc" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.866113 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f6e7467-e546-4563-bbe5-30d181f18349-config-data\") pod \"nova-cell0-conductor-db-sync-fmdjc\" (UID: \"5f6e7467-e546-4563-bbe5-30d181f18349\") " pod="openstack/nova-cell0-conductor-db-sync-fmdjc" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.869060 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gchdq\" (UniqueName: \"kubernetes.io/projected/5f6e7467-e546-4563-bbe5-30d181f18349-kube-api-access-gchdq\") pod \"nova-cell0-conductor-db-sync-fmdjc\" (UID: \"5f6e7467-e546-4563-bbe5-30d181f18349\") " pod="openstack/nova-cell0-conductor-db-sync-fmdjc" Oct 03 13:14:10 crc kubenswrapper[4578]: I1003 13:14:10.964161 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fmdjc" Oct 03 13:14:11 crc kubenswrapper[4578]: I1003 13:14:11.221005 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 03 13:14:11 crc kubenswrapper[4578]: I1003 13:14:11.221040 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 03 13:14:11 crc kubenswrapper[4578]: I1003 13:14:11.508832 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fmdjc"] Oct 03 13:14:12 crc kubenswrapper[4578]: I1003 13:14:12.227003 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fmdjc" event={"ID":"5f6e7467-e546-4563-bbe5-30d181f18349","Type":"ContainerStarted","Data":"f7f2d30f044ae3582fa8b8b51d55c72c8b942bba62861b166b94e648fe32333b"} Oct 03 13:14:13 crc kubenswrapper[4578]: I1003 13:14:13.568925 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 03 13:14:13 crc kubenswrapper[4578]: I1003 13:14:13.569327 4578 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 03 13:14:13 crc kubenswrapper[4578]: I1003 13:14:13.588246 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 03 13:14:16 crc kubenswrapper[4578]: I1003 13:14:16.816832 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7b58cf9568-f8nng" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 03 13:14:16 crc kubenswrapper[4578]: I1003 13:14:16.949059 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-68684899bd-2cwll" podUID="d49d623e-2cd6-4786-b1da-34b78b89a134" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 03 13:14:22 crc kubenswrapper[4578]: I1003 13:14:22.328407 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fmdjc" event={"ID":"5f6e7467-e546-4563-bbe5-30d181f18349","Type":"ContainerStarted","Data":"33b8b84c54d509a4a1b32d0b2f39a2180de18bdc686c351fb9eb8af094028901"} Oct 03 13:14:22 crc kubenswrapper[4578]: I1003 13:14:22.350610 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-fmdjc" podStartSLOduration=2.222700158 podStartE2EDuration="12.350587306s" podCreationTimestamp="2025-10-03 13:14:10 +0000 UTC" firstStartedPulling="2025-10-03 13:14:11.532131038 +0000 UTC m=+1387.330603222" lastFinishedPulling="2025-10-03 13:14:21.660018186 +0000 UTC m=+1397.458490370" observedRunningTime="2025-10-03 13:14:22.342033558 +0000 UTC m=+1398.140505772" watchObservedRunningTime="2025-10-03 13:14:22.350587306 +0000 UTC m=+1398.149059500" Oct 03 13:14:23 crc kubenswrapper[4578]: I1003 13:14:23.108913 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.393093 4578 generic.go:334] "Generic (PLEG): container finished" podID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerID="9c2664687a2e5ce8899e53ba9c5dfffce4a6572d01dd5e672b8a756e71a4722a" exitCode=137 Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.393311 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a925d503-688a-4a63-a6d5-1533a9a22aac","Type":"ContainerDied","Data":"9c2664687a2e5ce8899e53ba9c5dfffce4a6572d01dd5e672b8a756e71a4722a"} Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.393592 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a925d503-688a-4a63-a6d5-1533a9a22aac","Type":"ContainerDied","Data":"dcbaf256bf4768b662fb7ce87eaa185bef1264b7eb65a59cc2eeeba65ba377f3"} Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.393608 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcbaf256bf4768b662fb7ce87eaa185bef1264b7eb65a59cc2eeeba65ba377f3" Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.397737 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.491129 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-sg-core-conf-yaml\") pod \"a925d503-688a-4a63-a6d5-1533a9a22aac\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.491198 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-scripts\") pod \"a925d503-688a-4a63-a6d5-1533a9a22aac\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.491345 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-config-data\") pod \"a925d503-688a-4a63-a6d5-1533a9a22aac\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.491366 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a925d503-688a-4a63-a6d5-1533a9a22aac-log-httpd\") pod \"a925d503-688a-4a63-a6d5-1533a9a22aac\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.491394 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-combined-ca-bundle\") pod \"a925d503-688a-4a63-a6d5-1533a9a22aac\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.491429 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a925d503-688a-4a63-a6d5-1533a9a22aac-run-httpd\") pod \"a925d503-688a-4a63-a6d5-1533a9a22aac\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.491469 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdwd4\" (UniqueName: \"kubernetes.io/projected/a925d503-688a-4a63-a6d5-1533a9a22aac-kube-api-access-qdwd4\") pod \"a925d503-688a-4a63-a6d5-1533a9a22aac\" (UID: \"a925d503-688a-4a63-a6d5-1533a9a22aac\") " Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.492232 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a925d503-688a-4a63-a6d5-1533a9a22aac-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a925d503-688a-4a63-a6d5-1533a9a22aac" (UID: "a925d503-688a-4a63-a6d5-1533a9a22aac"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.492962 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a925d503-688a-4a63-a6d5-1533a9a22aac-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a925d503-688a-4a63-a6d5-1533a9a22aac" (UID: "a925d503-688a-4a63-a6d5-1533a9a22aac"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.509795 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a925d503-688a-4a63-a6d5-1533a9a22aac-kube-api-access-qdwd4" (OuterVolumeSpecName: "kube-api-access-qdwd4") pod "a925d503-688a-4a63-a6d5-1533a9a22aac" (UID: "a925d503-688a-4a63-a6d5-1533a9a22aac"). InnerVolumeSpecName "kube-api-access-qdwd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.527082 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-scripts" (OuterVolumeSpecName: "scripts") pod "a925d503-688a-4a63-a6d5-1533a9a22aac" (UID: "a925d503-688a-4a63-a6d5-1533a9a22aac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.546237 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a925d503-688a-4a63-a6d5-1533a9a22aac" (UID: "a925d503-688a-4a63-a6d5-1533a9a22aac"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.589973 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a925d503-688a-4a63-a6d5-1533a9a22aac" (UID: "a925d503-688a-4a63-a6d5-1533a9a22aac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.594498 4578 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a925d503-688a-4a63-a6d5-1533a9a22aac-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.594528 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdwd4\" (UniqueName: \"kubernetes.io/projected/a925d503-688a-4a63-a6d5-1533a9a22aac-kube-api-access-qdwd4\") on node \"crc\" DevicePath \"\"" Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.594539 4578 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.594976 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.594991 4578 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a925d503-688a-4a63-a6d5-1533a9a22aac-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.594999 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.630886 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-config-data" (OuterVolumeSpecName: "config-data") pod "a925d503-688a-4a63-a6d5-1533a9a22aac" (UID: "a925d503-688a-4a63-a6d5-1533a9a22aac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:14:28 crc kubenswrapper[4578]: I1003 13:14:28.697024 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a925d503-688a-4a63-a6d5-1533a9a22aac-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.049183 4578 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podcafa819d-515a-4e7a-8f97-e97522c4b31c"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podcafa819d-515a-4e7a-8f97-e97522c4b31c] : Timed out while waiting for systemd to remove kubepods-besteffort-podcafa819d_515a_4e7a_8f97_e97522c4b31c.slice" Oct 03 13:14:29 crc kubenswrapper[4578]: E1003 13:14:29.049235 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort podcafa819d-515a-4e7a-8f97-e97522c4b31c] : unable to destroy cgroup paths for cgroup [kubepods besteffort podcafa819d-515a-4e7a-8f97-e97522c4b31c] : Timed out while waiting for systemd to remove kubepods-besteffort-podcafa819d_515a_4e7a_8f97_e97522c4b31c.slice" pod="openstack/nova-cell0-db-create-56phj" podUID="cafa819d-515a-4e7a-8f97-e97522c4b31c" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.400915 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.400915 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-56phj" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.422519 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.431672 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.455593 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.455915 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:14:29 crc kubenswrapper[4578]: E1003 13:14:29.456444 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerName="ceilometer-notification-agent" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.456477 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerName="ceilometer-notification-agent" Oct 03 13:14:29 crc kubenswrapper[4578]: E1003 13:14:29.456492 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerName="proxy-httpd" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.456499 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerName="proxy-httpd" Oct 03 13:14:29 crc kubenswrapper[4578]: E1003 13:14:29.456540 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerName="sg-core" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.456548 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerName="sg-core" Oct 03 13:14:29 crc kubenswrapper[4578]: E1003 13:14:29.456564 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerName="ceilometer-central-agent" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.456571 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerName="ceilometer-central-agent" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.456872 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerName="ceilometer-notification-agent" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.456899 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerName="ceilometer-central-agent" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.456921 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerName="proxy-httpd" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.456941 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="a925d503-688a-4a63-a6d5-1533a9a22aac" containerName="sg-core" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.461670 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.475750 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.477786 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.477871 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.488707 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.636402 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.636501 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.636678 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c345497-9620-482e-abfa-868112c6e1e6-log-httpd\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.636702 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-scripts\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.636762 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-config-data\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.636853 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c345497-9620-482e-abfa-868112c6e1e6-run-httpd\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.636874 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s57c5\" (UniqueName: \"kubernetes.io/projected/4c345497-9620-482e-abfa-868112c6e1e6-kube-api-access-s57c5\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.738785 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.738940 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c345497-9620-482e-abfa-868112c6e1e6-log-httpd\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.738969 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-scripts\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.739015 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-config-data\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.739041 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c345497-9620-482e-abfa-868112c6e1e6-run-httpd\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.739065 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s57c5\" (UniqueName: \"kubernetes.io/projected/4c345497-9620-482e-abfa-868112c6e1e6-kube-api-access-s57c5\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.739118 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.740418 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c345497-9620-482e-abfa-868112c6e1e6-run-httpd\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.740884 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c345497-9620-482e-abfa-868112c6e1e6-log-httpd\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.745511 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.764342 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-scripts\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.764828 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.772435 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-config-data\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.775344 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s57c5\" (UniqueName: \"kubernetes.io/projected/4c345497-9620-482e-abfa-868112c6e1e6-kube-api-access-s57c5\") pod \"ceilometer-0\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " pod="openstack/ceilometer-0" Oct 03 13:14:29 crc kubenswrapper[4578]: I1003 13:14:29.784326 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:14:30 crc kubenswrapper[4578]: I1003 13:14:30.364822 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:14:30 crc kubenswrapper[4578]: I1003 13:14:30.385136 4578 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 13:14:30 crc kubenswrapper[4578]: I1003 13:14:30.412815 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c345497-9620-482e-abfa-868112c6e1e6","Type":"ContainerStarted","Data":"027771361b2a1d950bc6dc3775bf77cd0f477140f9338a9d913dc66aee287485"} Oct 03 13:14:30 crc kubenswrapper[4578]: I1003 13:14:30.923860 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a925d503-688a-4a63-a6d5-1533a9a22aac" path="/var/lib/kubelet/pods/a925d503-688a-4a63-a6d5-1533a9a22aac/volumes" Oct 03 13:14:31 crc kubenswrapper[4578]: I1003 13:14:31.423354 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c345497-9620-482e-abfa-868112c6e1e6","Type":"ContainerStarted","Data":"2fa6108292cf4712139d664537bc28e717b7be6deb73829c486d76891ae137f1"} Oct 03 13:14:31 crc kubenswrapper[4578]: I1003 13:14:31.566720 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:14:31 crc kubenswrapper[4578]: I1003 13:14:31.610103 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-68684899bd-2cwll" Oct 03 13:14:31 crc kubenswrapper[4578]: I1003 13:14:31.693772 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7b58cf9568-f8nng"] Oct 03 13:14:32 crc kubenswrapper[4578]: I1003 13:14:32.433269 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c345497-9620-482e-abfa-868112c6e1e6","Type":"ContainerStarted","Data":"2903e87c0cb090ddcda043a9f1ec22e7ed66c8e3fbf355cc766c6bc247368eae"} Oct 03 13:14:32 crc kubenswrapper[4578]: I1003 13:14:32.433376 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7b58cf9568-f8nng" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerName="horizon-log" containerID="cri-o://de051c4c2fdbf56aef0e53a639dcd8953e6218bb9bdef3f691e52909ccf616ec" gracePeriod=30 Oct 03 13:14:32 crc kubenswrapper[4578]: I1003 13:14:32.433405 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7b58cf9568-f8nng" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerName="horizon" containerID="cri-o://de8222a725afcbc467d7921332c4ab3214da8e901c323972f2737910bf0d32c3" gracePeriod=30 Oct 03 13:14:33 crc kubenswrapper[4578]: I1003 13:14:33.469492 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c345497-9620-482e-abfa-868112c6e1e6","Type":"ContainerStarted","Data":"1dc12dff3e12a4279cd27e0b3cfcb19a393615a985c7380d38cf793c967c9e1a"} Oct 03 13:14:33 crc kubenswrapper[4578]: I1003 13:14:33.471754 4578 generic.go:334] "Generic (PLEG): container finished" podID="5f6e7467-e546-4563-bbe5-30d181f18349" containerID="33b8b84c54d509a4a1b32d0b2f39a2180de18bdc686c351fb9eb8af094028901" exitCode=0 Oct 03 13:14:33 crc kubenswrapper[4578]: I1003 13:14:33.471790 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fmdjc" event={"ID":"5f6e7467-e546-4563-bbe5-30d181f18349","Type":"ContainerDied","Data":"33b8b84c54d509a4a1b32d0b2f39a2180de18bdc686c351fb9eb8af094028901"} Oct 03 13:14:34 crc kubenswrapper[4578]: I1003 13:14:34.482539 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c345497-9620-482e-abfa-868112c6e1e6","Type":"ContainerStarted","Data":"9570a50ac3ed7d4fc6eebdda73ed2bba5de9e90a534b24b6494319a43857cf9b"} Oct 03 13:14:34 crc kubenswrapper[4578]: I1003 13:14:34.483077 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 13:14:34 crc kubenswrapper[4578]: I1003 13:14:34.506319 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.271194922 podStartE2EDuration="5.506303275s" podCreationTimestamp="2025-10-03 13:14:29 +0000 UTC" firstStartedPulling="2025-10-03 13:14:30.384941224 +0000 UTC m=+1406.183413398" lastFinishedPulling="2025-10-03 13:14:33.620049567 +0000 UTC m=+1409.418521751" observedRunningTime="2025-10-03 13:14:34.504206759 +0000 UTC m=+1410.302678943" watchObservedRunningTime="2025-10-03 13:14:34.506303275 +0000 UTC m=+1410.304775459" Oct 03 13:14:34 crc kubenswrapper[4578]: I1003 13:14:34.830269 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fmdjc" Oct 03 13:14:34 crc kubenswrapper[4578]: I1003 13:14:34.955777 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f6e7467-e546-4563-bbe5-30d181f18349-combined-ca-bundle\") pod \"5f6e7467-e546-4563-bbe5-30d181f18349\" (UID: \"5f6e7467-e546-4563-bbe5-30d181f18349\") " Oct 03 13:14:34 crc kubenswrapper[4578]: I1003 13:14:34.955830 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f6e7467-e546-4563-bbe5-30d181f18349-scripts\") pod \"5f6e7467-e546-4563-bbe5-30d181f18349\" (UID: \"5f6e7467-e546-4563-bbe5-30d181f18349\") " Oct 03 13:14:34 crc kubenswrapper[4578]: I1003 13:14:34.955862 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f6e7467-e546-4563-bbe5-30d181f18349-config-data\") pod \"5f6e7467-e546-4563-bbe5-30d181f18349\" (UID: \"5f6e7467-e546-4563-bbe5-30d181f18349\") " Oct 03 13:14:34 crc kubenswrapper[4578]: I1003 13:14:34.955918 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gchdq\" (UniqueName: \"kubernetes.io/projected/5f6e7467-e546-4563-bbe5-30d181f18349-kube-api-access-gchdq\") pod \"5f6e7467-e546-4563-bbe5-30d181f18349\" (UID: \"5f6e7467-e546-4563-bbe5-30d181f18349\") " Oct 03 13:14:34 crc kubenswrapper[4578]: I1003 13:14:34.961206 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f6e7467-e546-4563-bbe5-30d181f18349-kube-api-access-gchdq" (OuterVolumeSpecName: "kube-api-access-gchdq") pod "5f6e7467-e546-4563-bbe5-30d181f18349" (UID: "5f6e7467-e546-4563-bbe5-30d181f18349"). InnerVolumeSpecName "kube-api-access-gchdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:14:34 crc kubenswrapper[4578]: I1003 13:14:34.965538 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f6e7467-e546-4563-bbe5-30d181f18349-scripts" (OuterVolumeSpecName: "scripts") pod "5f6e7467-e546-4563-bbe5-30d181f18349" (UID: "5f6e7467-e546-4563-bbe5-30d181f18349"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:14:34 crc kubenswrapper[4578]: I1003 13:14:34.990272 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f6e7467-e546-4563-bbe5-30d181f18349-config-data" (OuterVolumeSpecName: "config-data") pod "5f6e7467-e546-4563-bbe5-30d181f18349" (UID: "5f6e7467-e546-4563-bbe5-30d181f18349"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.006322 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f6e7467-e546-4563-bbe5-30d181f18349-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5f6e7467-e546-4563-bbe5-30d181f18349" (UID: "5f6e7467-e546-4563-bbe5-30d181f18349"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.058099 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f6e7467-e546-4563-bbe5-30d181f18349-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.058140 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f6e7467-e546-4563-bbe5-30d181f18349-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.058312 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f6e7467-e546-4563-bbe5-30d181f18349-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.058328 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gchdq\" (UniqueName: \"kubernetes.io/projected/5f6e7467-e546-4563-bbe5-30d181f18349-kube-api-access-gchdq\") on node \"crc\" DevicePath \"\"" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.491568 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fmdjc" event={"ID":"5f6e7467-e546-4563-bbe5-30d181f18349","Type":"ContainerDied","Data":"f7f2d30f044ae3582fa8b8b51d55c72c8b942bba62861b166b94e648fe32333b"} Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.491880 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7f2d30f044ae3582fa8b8b51d55c72c8b942bba62861b166b94e648fe32333b" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.491576 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fmdjc" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.600113 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 13:14:35 crc kubenswrapper[4578]: E1003 13:14:35.600581 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f6e7467-e546-4563-bbe5-30d181f18349" containerName="nova-cell0-conductor-db-sync" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.600601 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f6e7467-e546-4563-bbe5-30d181f18349" containerName="nova-cell0-conductor-db-sync" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.600891 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f6e7467-e546-4563-bbe5-30d181f18349" containerName="nova-cell0-conductor-db-sync" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.601623 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.604096 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.606017 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-8dm6s" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.621538 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.675556 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj97w\" (UniqueName: \"kubernetes.io/projected/cf61a63d-477b-4e6d-9946-fd268045d303-kube-api-access-vj97w\") pod \"nova-cell0-conductor-0\" (UID: \"cf61a63d-477b-4e6d-9946-fd268045d303\") " pod="openstack/nova-cell0-conductor-0" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.675708 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf61a63d-477b-4e6d-9946-fd268045d303-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"cf61a63d-477b-4e6d-9946-fd268045d303\") " pod="openstack/nova-cell0-conductor-0" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.675730 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf61a63d-477b-4e6d-9946-fd268045d303-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"cf61a63d-477b-4e6d-9946-fd268045d303\") " pod="openstack/nova-cell0-conductor-0" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.778253 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf61a63d-477b-4e6d-9946-fd268045d303-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"cf61a63d-477b-4e6d-9946-fd268045d303\") " pod="openstack/nova-cell0-conductor-0" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.778644 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf61a63d-477b-4e6d-9946-fd268045d303-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"cf61a63d-477b-4e6d-9946-fd268045d303\") " pod="openstack/nova-cell0-conductor-0" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.778814 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj97w\" (UniqueName: \"kubernetes.io/projected/cf61a63d-477b-4e6d-9946-fd268045d303-kube-api-access-vj97w\") pod \"nova-cell0-conductor-0\" (UID: \"cf61a63d-477b-4e6d-9946-fd268045d303\") " pod="openstack/nova-cell0-conductor-0" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.784543 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf61a63d-477b-4e6d-9946-fd268045d303-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"cf61a63d-477b-4e6d-9946-fd268045d303\") " pod="openstack/nova-cell0-conductor-0" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.785545 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf61a63d-477b-4e6d-9946-fd268045d303-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"cf61a63d-477b-4e6d-9946-fd268045d303\") " pod="openstack/nova-cell0-conductor-0" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.795692 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj97w\" (UniqueName: \"kubernetes.io/projected/cf61a63d-477b-4e6d-9946-fd268045d303-kube-api-access-vj97w\") pod \"nova-cell0-conductor-0\" (UID: \"cf61a63d-477b-4e6d-9946-fd268045d303\") " pod="openstack/nova-cell0-conductor-0" Oct 03 13:14:35 crc kubenswrapper[4578]: I1003 13:14:35.926558 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 03 13:14:36 crc kubenswrapper[4578]: I1003 13:14:36.368470 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 03 13:14:36 crc kubenswrapper[4578]: W1003 13:14:36.379940 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf61a63d_477b_4e6d_9946_fd268045d303.slice/crio-7853fb5be135bb9d672dfd2aae757e985fc881b8485a80a3f819777815dae4cd WatchSource:0}: Error finding container 7853fb5be135bb9d672dfd2aae757e985fc881b8485a80a3f819777815dae4cd: Status 404 returned error can't find the container with id 7853fb5be135bb9d672dfd2aae757e985fc881b8485a80a3f819777815dae4cd Oct 03 13:14:36 crc kubenswrapper[4578]: I1003 13:14:36.503544 4578 generic.go:334] "Generic (PLEG): container finished" podID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerID="de8222a725afcbc467d7921332c4ab3214da8e901c323972f2737910bf0d32c3" exitCode=0 Oct 03 13:14:36 crc kubenswrapper[4578]: I1003 13:14:36.503835 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b58cf9568-f8nng" event={"ID":"a54aea6a-c051-4832-b764-3ef53a9cc3a1","Type":"ContainerDied","Data":"de8222a725afcbc467d7921332c4ab3214da8e901c323972f2737910bf0d32c3"} Oct 03 13:14:36 crc kubenswrapper[4578]: I1003 13:14:36.503867 4578 scope.go:117] "RemoveContainer" containerID="033a127d7b3dc6a5469b25fc2682629ad13d3e1ed8dd904aad8edc6112ea940b" Oct 03 13:14:36 crc kubenswrapper[4578]: I1003 13:14:36.505658 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"cf61a63d-477b-4e6d-9946-fd268045d303","Type":"ContainerStarted","Data":"7853fb5be135bb9d672dfd2aae757e985fc881b8485a80a3f819777815dae4cd"} Oct 03 13:14:36 crc kubenswrapper[4578]: I1003 13:14:36.780715 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7b58cf9568-f8nng" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 03 13:14:37 crc kubenswrapper[4578]: I1003 13:14:37.517571 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"cf61a63d-477b-4e6d-9946-fd268045d303","Type":"ContainerStarted","Data":"08372aae4a08530c54fbea0e05128a77fd1d70597c10cc37f465137e4487c316"} Oct 03 13:14:37 crc kubenswrapper[4578]: I1003 13:14:37.519159 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 03 13:14:37 crc kubenswrapper[4578]: I1003 13:14:37.539703 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.539685521 podStartE2EDuration="2.539685521s" podCreationTimestamp="2025-10-03 13:14:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:14:37.538853385 +0000 UTC m=+1413.337325579" watchObservedRunningTime="2025-10-03 13:14:37.539685521 +0000 UTC m=+1413.338157705" Oct 03 13:14:45 crc kubenswrapper[4578]: I1003 13:14:45.955507 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.453489 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-nvwxn"] Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.455419 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-nvwxn" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.460142 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.460593 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.469791 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-nvwxn"] Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.671710 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4304b326-aece-4557-904b-31f2fbab5c1f-config-data\") pod \"nova-cell0-cell-mapping-nvwxn\" (UID: \"4304b326-aece-4557-904b-31f2fbab5c1f\") " pod="openstack/nova-cell0-cell-mapping-nvwxn" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.671811 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4304b326-aece-4557-904b-31f2fbab5c1f-scripts\") pod \"nova-cell0-cell-mapping-nvwxn\" (UID: \"4304b326-aece-4557-904b-31f2fbab5c1f\") " pod="openstack/nova-cell0-cell-mapping-nvwxn" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.671834 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4304b326-aece-4557-904b-31f2fbab5c1f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-nvwxn\" (UID: \"4304b326-aece-4557-904b-31f2fbab5c1f\") " pod="openstack/nova-cell0-cell-mapping-nvwxn" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.671897 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvd5x\" (UniqueName: \"kubernetes.io/projected/4304b326-aece-4557-904b-31f2fbab5c1f-kube-api-access-fvd5x\") pod \"nova-cell0-cell-mapping-nvwxn\" (UID: \"4304b326-aece-4557-904b-31f2fbab5c1f\") " pod="openstack/nova-cell0-cell-mapping-nvwxn" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.674590 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.678497 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.685092 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.720517 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.773639 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvd5x\" (UniqueName: \"kubernetes.io/projected/4304b326-aece-4557-904b-31f2fbab5c1f-kube-api-access-fvd5x\") pod \"nova-cell0-cell-mapping-nvwxn\" (UID: \"4304b326-aece-4557-904b-31f2fbab5c1f\") " pod="openstack/nova-cell0-cell-mapping-nvwxn" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.774003 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4304b326-aece-4557-904b-31f2fbab5c1f-config-data\") pod \"nova-cell0-cell-mapping-nvwxn\" (UID: \"4304b326-aece-4557-904b-31f2fbab5c1f\") " pod="openstack/nova-cell0-cell-mapping-nvwxn" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.774089 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4304b326-aece-4557-904b-31f2fbab5c1f-scripts\") pod \"nova-cell0-cell-mapping-nvwxn\" (UID: \"4304b326-aece-4557-904b-31f2fbab5c1f\") " pod="openstack/nova-cell0-cell-mapping-nvwxn" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.774113 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4304b326-aece-4557-904b-31f2fbab5c1f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-nvwxn\" (UID: \"4304b326-aece-4557-904b-31f2fbab5c1f\") " pod="openstack/nova-cell0-cell-mapping-nvwxn" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.789035 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7b58cf9568-f8nng" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.826565 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4304b326-aece-4557-904b-31f2fbab5c1f-config-data\") pod \"nova-cell0-cell-mapping-nvwxn\" (UID: \"4304b326-aece-4557-904b-31f2fbab5c1f\") " pod="openstack/nova-cell0-cell-mapping-nvwxn" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.831565 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4304b326-aece-4557-904b-31f2fbab5c1f-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-nvwxn\" (UID: \"4304b326-aece-4557-904b-31f2fbab5c1f\") " pod="openstack/nova-cell0-cell-mapping-nvwxn" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.831924 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4304b326-aece-4557-904b-31f2fbab5c1f-scripts\") pod \"nova-cell0-cell-mapping-nvwxn\" (UID: \"4304b326-aece-4557-904b-31f2fbab5c1f\") " pod="openstack/nova-cell0-cell-mapping-nvwxn" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.836283 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvd5x\" (UniqueName: \"kubernetes.io/projected/4304b326-aece-4557-904b-31f2fbab5c1f-kube-api-access-fvd5x\") pod \"nova-cell0-cell-mapping-nvwxn\" (UID: \"4304b326-aece-4557-904b-31f2fbab5c1f\") " pod="openstack/nova-cell0-cell-mapping-nvwxn" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.875157 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.876074 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b\") " pod="openstack/nova-api-0" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.876261 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-logs\") pod \"nova-api-0\" (UID: \"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b\") " pod="openstack/nova-api-0" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.876365 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bclk\" (UniqueName: \"kubernetes.io/projected/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-kube-api-access-7bclk\") pod \"nova-api-0\" (UID: \"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b\") " pod="openstack/nova-api-0" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.876467 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-config-data\") pod \"nova-api-0\" (UID: \"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b\") " pod="openstack/nova-api-0" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.879515 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.897477 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.901462 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.983662 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbaa7077-fdc1-49d9-a05b-213ed1f50138-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fbaa7077-fdc1-49d9-a05b-213ed1f50138\") " pod="openstack/nova-scheduler-0" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.983753 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b\") " pod="openstack/nova-api-0" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.983782 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbaa7077-fdc1-49d9-a05b-213ed1f50138-config-data\") pod \"nova-scheduler-0\" (UID: \"fbaa7077-fdc1-49d9-a05b-213ed1f50138\") " pod="openstack/nova-scheduler-0" Oct 03 13:14:46 crc kubenswrapper[4578]: I1003 13:14:46.983801 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbhr4\" (UniqueName: \"kubernetes.io/projected/fbaa7077-fdc1-49d9-a05b-213ed1f50138-kube-api-access-cbhr4\") pod \"nova-scheduler-0\" (UID: \"fbaa7077-fdc1-49d9-a05b-213ed1f50138\") " pod="openstack/nova-scheduler-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:46.983846 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-logs\") pod \"nova-api-0\" (UID: \"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b\") " pod="openstack/nova-api-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:46.983875 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bclk\" (UniqueName: \"kubernetes.io/projected/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-kube-api-access-7bclk\") pod \"nova-api-0\" (UID: \"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b\") " pod="openstack/nova-api-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:46.983892 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-config-data\") pod \"nova-api-0\" (UID: \"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b\") " pod="openstack/nova-api-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:46.985777 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-logs\") pod \"nova-api-0\" (UID: \"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b\") " pod="openstack/nova-api-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:46.997423 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b\") " pod="openstack/nova-api-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.003320 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-config-data\") pod \"nova-api-0\" (UID: \"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b\") " pod="openstack/nova-api-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.026340 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bclk\" (UniqueName: \"kubernetes.io/projected/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-kube-api-access-7bclk\") pod \"nova-api-0\" (UID: \"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b\") " pod="openstack/nova-api-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.080324 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-nvwxn" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.088788 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbaa7077-fdc1-49d9-a05b-213ed1f50138-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fbaa7077-fdc1-49d9-a05b-213ed1f50138\") " pod="openstack/nova-scheduler-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.088879 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbaa7077-fdc1-49d9-a05b-213ed1f50138-config-data\") pod \"nova-scheduler-0\" (UID: \"fbaa7077-fdc1-49d9-a05b-213ed1f50138\") " pod="openstack/nova-scheduler-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.088899 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbhr4\" (UniqueName: \"kubernetes.io/projected/fbaa7077-fdc1-49d9-a05b-213ed1f50138-kube-api-access-cbhr4\") pod \"nova-scheduler-0\" (UID: \"fbaa7077-fdc1-49d9-a05b-213ed1f50138\") " pod="openstack/nova-scheduler-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.112801 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbaa7077-fdc1-49d9-a05b-213ed1f50138-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fbaa7077-fdc1-49d9-a05b-213ed1f50138\") " pod="openstack/nova-scheduler-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.121402 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbaa7077-fdc1-49d9-a05b-213ed1f50138-config-data\") pod \"nova-scheduler-0\" (UID: \"fbaa7077-fdc1-49d9-a05b-213ed1f50138\") " pod="openstack/nova-scheduler-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.151850 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbhr4\" (UniqueName: \"kubernetes.io/projected/fbaa7077-fdc1-49d9-a05b-213ed1f50138-kube-api-access-cbhr4\") pod \"nova-scheduler-0\" (UID: \"fbaa7077-fdc1-49d9-a05b-213ed1f50138\") " pod="openstack/nova-scheduler-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.319453 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.425491 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.429731 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.430291 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.430316 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.431481 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.431524 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-g7kjh"] Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.433306 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.433525 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.433663 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.433531 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-g7kjh"] Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.468768 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.488375 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.535746 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-config\") pod \"dnsmasq-dns-845d6d6f59-g7kjh\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.535793 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-g7kjh\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.535909 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.535934 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/241fb208-e394-48e3-acf7-5ec5c90ee78a-config-data\") pod \"nova-metadata-0\" (UID: \"241fb208-e394-48e3-acf7-5ec5c90ee78a\") " pod="openstack/nova-metadata-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.535957 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdwcg\" (UniqueName: \"kubernetes.io/projected/7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1-kube-api-access-cdwcg\") pod \"nova-cell1-novncproxy-0\" (UID: \"7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.535975 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.536009 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g4sj\" (UniqueName: \"kubernetes.io/projected/81676263-327d-4d19-8673-a44d5e576430-kube-api-access-6g4sj\") pod \"dnsmasq-dns-845d6d6f59-g7kjh\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.536044 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z767w\" (UniqueName: \"kubernetes.io/projected/241fb208-e394-48e3-acf7-5ec5c90ee78a-kube-api-access-z767w\") pod \"nova-metadata-0\" (UID: \"241fb208-e394-48e3-acf7-5ec5c90ee78a\") " pod="openstack/nova-metadata-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.536067 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-g7kjh\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.536111 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/241fb208-e394-48e3-acf7-5ec5c90ee78a-logs\") pod \"nova-metadata-0\" (UID: \"241fb208-e394-48e3-acf7-5ec5c90ee78a\") " pod="openstack/nova-metadata-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.536151 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-g7kjh\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.536198 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-g7kjh\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.536239 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241fb208-e394-48e3-acf7-5ec5c90ee78a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"241fb208-e394-48e3-acf7-5ec5c90ee78a\") " pod="openstack/nova-metadata-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.640914 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.640949 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/241fb208-e394-48e3-acf7-5ec5c90ee78a-config-data\") pod \"nova-metadata-0\" (UID: \"241fb208-e394-48e3-acf7-5ec5c90ee78a\") " pod="openstack/nova-metadata-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.640967 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdwcg\" (UniqueName: \"kubernetes.io/projected/7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1-kube-api-access-cdwcg\") pod \"nova-cell1-novncproxy-0\" (UID: \"7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.640983 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.641007 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g4sj\" (UniqueName: \"kubernetes.io/projected/81676263-327d-4d19-8673-a44d5e576430-kube-api-access-6g4sj\") pod \"dnsmasq-dns-845d6d6f59-g7kjh\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.641031 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z767w\" (UniqueName: \"kubernetes.io/projected/241fb208-e394-48e3-acf7-5ec5c90ee78a-kube-api-access-z767w\") pod \"nova-metadata-0\" (UID: \"241fb208-e394-48e3-acf7-5ec5c90ee78a\") " pod="openstack/nova-metadata-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.641047 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-g7kjh\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.641069 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/241fb208-e394-48e3-acf7-5ec5c90ee78a-logs\") pod \"nova-metadata-0\" (UID: \"241fb208-e394-48e3-acf7-5ec5c90ee78a\") " pod="openstack/nova-metadata-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.641098 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-g7kjh\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.641120 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-g7kjh\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.641139 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241fb208-e394-48e3-acf7-5ec5c90ee78a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"241fb208-e394-48e3-acf7-5ec5c90ee78a\") " pod="openstack/nova-metadata-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.641187 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-config\") pod \"dnsmasq-dns-845d6d6f59-g7kjh\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.641255 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-g7kjh\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.658477 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-g7kjh\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.670499 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-g7kjh\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.673167 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.674479 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-g7kjh\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.676908 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/241fb208-e394-48e3-acf7-5ec5c90ee78a-logs\") pod \"nova-metadata-0\" (UID: \"241fb208-e394-48e3-acf7-5ec5c90ee78a\") " pod="openstack/nova-metadata-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.684310 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-config\") pod \"dnsmasq-dns-845d6d6f59-g7kjh\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.701371 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-g7kjh\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.729317 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241fb208-e394-48e3-acf7-5ec5c90ee78a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"241fb208-e394-48e3-acf7-5ec5c90ee78a\") " pod="openstack/nova-metadata-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.748311 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/241fb208-e394-48e3-acf7-5ec5c90ee78a-config-data\") pod \"nova-metadata-0\" (UID: \"241fb208-e394-48e3-acf7-5ec5c90ee78a\") " pod="openstack/nova-metadata-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.762258 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z767w\" (UniqueName: \"kubernetes.io/projected/241fb208-e394-48e3-acf7-5ec5c90ee78a-kube-api-access-z767w\") pod \"nova-metadata-0\" (UID: \"241fb208-e394-48e3-acf7-5ec5c90ee78a\") " pod="openstack/nova-metadata-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.762437 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g4sj\" (UniqueName: \"kubernetes.io/projected/81676263-327d-4d19-8673-a44d5e576430-kube-api-access-6g4sj\") pod \"dnsmasq-dns-845d6d6f59-g7kjh\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.769382 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.816266 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdwcg\" (UniqueName: \"kubernetes.io/projected/7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1-kube-api-access-cdwcg\") pod \"nova-cell1-novncproxy-0\" (UID: \"7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.818018 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.844683 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.859043 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:47 crc kubenswrapper[4578]: I1003 13:14:47.864407 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-nvwxn"] Oct 03 13:14:48 crc kubenswrapper[4578]: I1003 13:14:48.360999 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:14:48 crc kubenswrapper[4578]: W1003 13:14:48.379746 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b1f1c2c_d1d8_4b74_a471_320b5a3bad6b.slice/crio-14c0258d47c7e75a1ec6d46f19fd09d746173ffea8ce801bc6c93f498813cd02 WatchSource:0}: Error finding container 14c0258d47c7e75a1ec6d46f19fd09d746173ffea8ce801bc6c93f498813cd02: Status 404 returned error can't find the container with id 14c0258d47c7e75a1ec6d46f19fd09d746173ffea8ce801bc6c93f498813cd02 Oct 03 13:14:48 crc kubenswrapper[4578]: I1003 13:14:48.585093 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:14:48 crc kubenswrapper[4578]: I1003 13:14:48.651544 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-nvwxn" event={"ID":"4304b326-aece-4557-904b-31f2fbab5c1f","Type":"ContainerStarted","Data":"2048be9c4225e7db96e0b1128fcfe04007983ee58ba570f83e79eebe69ed52b5"} Oct 03 13:14:48 crc kubenswrapper[4578]: I1003 13:14:48.651603 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-nvwxn" event={"ID":"4304b326-aece-4557-904b-31f2fbab5c1f","Type":"ContainerStarted","Data":"a89a20c7cf57d3138a3b21ef7ed631f359c8a3ac9fd0d96ccc8fd07973d457be"} Oct 03 13:14:48 crc kubenswrapper[4578]: I1003 13:14:48.659535 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b","Type":"ContainerStarted","Data":"14c0258d47c7e75a1ec6d46f19fd09d746173ffea8ce801bc6c93f498813cd02"} Oct 03 13:14:48 crc kubenswrapper[4578]: I1003 13:14:48.660781 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fbaa7077-fdc1-49d9-a05b-213ed1f50138","Type":"ContainerStarted","Data":"5b409837939e426cfc20b4bb0891b021475d1e86dd9859d47e76a8a11c4b9ab0"} Oct 03 13:14:48 crc kubenswrapper[4578]: I1003 13:14:48.687711 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-nvwxn" podStartSLOduration=2.687692003 podStartE2EDuration="2.687692003s" podCreationTimestamp="2025-10-03 13:14:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:14:48.676936255 +0000 UTC m=+1424.475408449" watchObservedRunningTime="2025-10-03 13:14:48.687692003 +0000 UTC m=+1424.486164187" Oct 03 13:14:48 crc kubenswrapper[4578]: I1003 13:14:48.779790 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-g7kjh"] Oct 03 13:14:48 crc kubenswrapper[4578]: I1003 13:14:48.898314 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:14:48 crc kubenswrapper[4578]: I1003 13:14:48.945739 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:14:48 crc kubenswrapper[4578]: W1003 13:14:48.955210 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod241fb208_e394_48e3_acf7_5ec5c90ee78a.slice/crio-1f24033bd84a2f7d1f850101a2765a6d3c9cfc6e0f7da4e07db534bdf2e96cce WatchSource:0}: Error finding container 1f24033bd84a2f7d1f850101a2765a6d3c9cfc6e0f7da4e07db534bdf2e96cce: Status 404 returned error can't find the container with id 1f24033bd84a2f7d1f850101a2765a6d3c9cfc6e0f7da4e07db534bdf2e96cce Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.055622 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-q4j79"] Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.056972 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-q4j79" Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.058708 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.059330 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.074340 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-q4j79"] Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.100687 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-q4j79\" (UID: \"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e\") " pod="openstack/nova-cell1-conductor-db-sync-q4j79" Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.100765 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-scripts\") pod \"nova-cell1-conductor-db-sync-q4j79\" (UID: \"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e\") " pod="openstack/nova-cell1-conductor-db-sync-q4j79" Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.100853 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-config-data\") pod \"nova-cell1-conductor-db-sync-q4j79\" (UID: \"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e\") " pod="openstack/nova-cell1-conductor-db-sync-q4j79" Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.100934 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrscn\" (UniqueName: \"kubernetes.io/projected/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-kube-api-access-wrscn\") pod \"nova-cell1-conductor-db-sync-q4j79\" (UID: \"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e\") " pod="openstack/nova-cell1-conductor-db-sync-q4j79" Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.203833 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-q4j79\" (UID: \"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e\") " pod="openstack/nova-cell1-conductor-db-sync-q4j79" Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.204951 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-scripts\") pod \"nova-cell1-conductor-db-sync-q4j79\" (UID: \"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e\") " pod="openstack/nova-cell1-conductor-db-sync-q4j79" Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.206441 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-config-data\") pod \"nova-cell1-conductor-db-sync-q4j79\" (UID: \"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e\") " pod="openstack/nova-cell1-conductor-db-sync-q4j79" Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.206554 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrscn\" (UniqueName: \"kubernetes.io/projected/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-kube-api-access-wrscn\") pod \"nova-cell1-conductor-db-sync-q4j79\" (UID: \"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e\") " pod="openstack/nova-cell1-conductor-db-sync-q4j79" Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.216111 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-scripts\") pod \"nova-cell1-conductor-db-sync-q4j79\" (UID: \"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e\") " pod="openstack/nova-cell1-conductor-db-sync-q4j79" Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.216909 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-config-data\") pod \"nova-cell1-conductor-db-sync-q4j79\" (UID: \"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e\") " pod="openstack/nova-cell1-conductor-db-sync-q4j79" Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.219312 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-q4j79\" (UID: \"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e\") " pod="openstack/nova-cell1-conductor-db-sync-q4j79" Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.228171 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrscn\" (UniqueName: \"kubernetes.io/projected/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-kube-api-access-wrscn\") pod \"nova-cell1-conductor-db-sync-q4j79\" (UID: \"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e\") " pod="openstack/nova-cell1-conductor-db-sync-q4j79" Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.401308 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-q4j79" Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.689499 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1","Type":"ContainerStarted","Data":"69c6104376243ce1eb0f873f51086ba110fed41fd227ba0492fa5f270ab9316f"} Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.691294 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"241fb208-e394-48e3-acf7-5ec5c90ee78a","Type":"ContainerStarted","Data":"1f24033bd84a2f7d1f850101a2765a6d3c9cfc6e0f7da4e07db534bdf2e96cce"} Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.692802 4578 generic.go:334] "Generic (PLEG): container finished" podID="81676263-327d-4d19-8673-a44d5e576430" containerID="92c3b9ede7c0673b1000809f1f762eab7b156b40fca03f0eb4d352abae8eaaa1" exitCode=0 Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.696493 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" event={"ID":"81676263-327d-4d19-8673-a44d5e576430","Type":"ContainerDied","Data":"92c3b9ede7c0673b1000809f1f762eab7b156b40fca03f0eb4d352abae8eaaa1"} Oct 03 13:14:49 crc kubenswrapper[4578]: I1003 13:14:49.696530 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" event={"ID":"81676263-327d-4d19-8673-a44d5e576430","Type":"ContainerStarted","Data":"9296bd0a9e59d957fe83229af5326fb0d254487b9278fc15f157d9dc6c38a77c"} Oct 03 13:14:50 crc kubenswrapper[4578]: I1003 13:14:50.032266 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-q4j79"] Oct 03 13:14:50 crc kubenswrapper[4578]: I1003 13:14:50.710600 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" event={"ID":"81676263-327d-4d19-8673-a44d5e576430","Type":"ContainerStarted","Data":"4b1b34611cc3274a88742b81445a6c0ca5c68f7a7c316eceb9dad6764ab43967"} Oct 03 13:14:50 crc kubenswrapper[4578]: I1003 13:14:50.711147 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:50 crc kubenswrapper[4578]: I1003 13:14:50.728309 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-q4j79" event={"ID":"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e","Type":"ContainerStarted","Data":"003a0c1928553820972cefea96c5d5af55961caad7bbb301c6b076d814788eec"} Oct 03 13:14:50 crc kubenswrapper[4578]: I1003 13:14:50.728346 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-q4j79" event={"ID":"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e","Type":"ContainerStarted","Data":"c51c0a4c7bfa9d9f5e7521e5ef3cbb5c80a0da9c759b258c2a536d5b410913bc"} Oct 03 13:14:50 crc kubenswrapper[4578]: I1003 13:14:50.754640 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" podStartSLOduration=3.7546082050000003 podStartE2EDuration="3.754608205s" podCreationTimestamp="2025-10-03 13:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:14:50.741367389 +0000 UTC m=+1426.539839573" watchObservedRunningTime="2025-10-03 13:14:50.754608205 +0000 UTC m=+1426.553080389" Oct 03 13:14:50 crc kubenswrapper[4578]: I1003 13:14:50.761403 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-q4j79" podStartSLOduration=1.761387048 podStartE2EDuration="1.761387048s" podCreationTimestamp="2025-10-03 13:14:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:14:50.756997611 +0000 UTC m=+1426.555469795" watchObservedRunningTime="2025-10-03 13:14:50.761387048 +0000 UTC m=+1426.559859232" Oct 03 13:14:51 crc kubenswrapper[4578]: I1003 13:14:51.251095 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:14:51 crc kubenswrapper[4578]: I1003 13:14:51.282735 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:14:53 crc kubenswrapper[4578]: I1003 13:14:53.771154 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fbaa7077-fdc1-49d9-a05b-213ed1f50138","Type":"ContainerStarted","Data":"546f2dbd38125228a514aeb67874f62b3a4dc0a03aa867e9f2b355d6007f3abb"} Oct 03 13:14:53 crc kubenswrapper[4578]: I1003 13:14:53.775446 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1","Type":"ContainerStarted","Data":"0d9425a149c0242a4eff40b79c7f0e23a726b37168550e73f9255f72a9e7c36c"} Oct 03 13:14:53 crc kubenswrapper[4578]: I1003 13:14:53.775513 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://0d9425a149c0242a4eff40b79c7f0e23a726b37168550e73f9255f72a9e7c36c" gracePeriod=30 Oct 03 13:14:53 crc kubenswrapper[4578]: I1003 13:14:53.780413 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"241fb208-e394-48e3-acf7-5ec5c90ee78a","Type":"ContainerStarted","Data":"88a63d3cbf44911329d647457f6985e80beb070f3eea72c841db8dca447a6e10"} Oct 03 13:14:53 crc kubenswrapper[4578]: I1003 13:14:53.780453 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"241fb208-e394-48e3-acf7-5ec5c90ee78a","Type":"ContainerStarted","Data":"03725b4989bcd78d56fe85a2a5a02fb423b38d935130808036ccd0327ba40e2e"} Oct 03 13:14:53 crc kubenswrapper[4578]: I1003 13:14:53.780498 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="241fb208-e394-48e3-acf7-5ec5c90ee78a" containerName="nova-metadata-metadata" containerID="cri-o://88a63d3cbf44911329d647457f6985e80beb070f3eea72c841db8dca447a6e10" gracePeriod=30 Oct 03 13:14:53 crc kubenswrapper[4578]: I1003 13:14:53.780491 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="241fb208-e394-48e3-acf7-5ec5c90ee78a" containerName="nova-metadata-log" containerID="cri-o://03725b4989bcd78d56fe85a2a5a02fb423b38d935130808036ccd0327ba40e2e" gracePeriod=30 Oct 03 13:14:53 crc kubenswrapper[4578]: I1003 13:14:53.800614 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.358749466 podStartE2EDuration="7.800595405s" podCreationTimestamp="2025-10-03 13:14:46 +0000 UTC" firstStartedPulling="2025-10-03 13:14:48.600805094 +0000 UTC m=+1424.399277278" lastFinishedPulling="2025-10-03 13:14:53.042651033 +0000 UTC m=+1428.841123217" observedRunningTime="2025-10-03 13:14:53.789667771 +0000 UTC m=+1429.588139955" watchObservedRunningTime="2025-10-03 13:14:53.800595405 +0000 UTC m=+1429.599067589" Oct 03 13:14:53 crc kubenswrapper[4578]: I1003 13:14:53.804769 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b","Type":"ContainerStarted","Data":"27198dd8e823529e9fd50a1f477c0d9e19e3f4106df8b6afdf29dbba0c13441c"} Oct 03 13:14:53 crc kubenswrapper[4578]: I1003 13:14:53.804842 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b","Type":"ContainerStarted","Data":"fbd39dc5b7e344c6b998d12c45c0c1cfa3556c8e4d54e1a7a57c01105dd03aba"} Oct 03 13:14:53 crc kubenswrapper[4578]: I1003 13:14:53.819499 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.69824783 podStartE2EDuration="7.819473358s" podCreationTimestamp="2025-10-03 13:14:46 +0000 UTC" firstStartedPulling="2025-10-03 13:14:48.918252466 +0000 UTC m=+1424.716724640" lastFinishedPulling="2025-10-03 13:14:53.039477984 +0000 UTC m=+1428.837950168" observedRunningTime="2025-10-03 13:14:53.809428852 +0000 UTC m=+1429.607901046" watchObservedRunningTime="2025-10-03 13:14:53.819473358 +0000 UTC m=+1429.617945542" Oct 03 13:14:53 crc kubenswrapper[4578]: I1003 13:14:53.836906 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.781600198 podStartE2EDuration="7.836882174s" podCreationTimestamp="2025-10-03 13:14:46 +0000 UTC" firstStartedPulling="2025-10-03 13:14:48.980818812 +0000 UTC m=+1424.779290996" lastFinishedPulling="2025-10-03 13:14:53.036100788 +0000 UTC m=+1428.834572972" observedRunningTime="2025-10-03 13:14:53.833118176 +0000 UTC m=+1429.631590360" watchObservedRunningTime="2025-10-03 13:14:53.836882174 +0000 UTC m=+1429.635354358" Oct 03 13:14:53 crc kubenswrapper[4578]: I1003 13:14:53.863618 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.22079894 podStartE2EDuration="7.863603424s" podCreationTimestamp="2025-10-03 13:14:46 +0000 UTC" firstStartedPulling="2025-10-03 13:14:48.396682021 +0000 UTC m=+1424.195154205" lastFinishedPulling="2025-10-03 13:14:53.039486505 +0000 UTC m=+1428.837958689" observedRunningTime="2025-10-03 13:14:53.862890452 +0000 UTC m=+1429.661362636" watchObservedRunningTime="2025-10-03 13:14:53.863603424 +0000 UTC m=+1429.662075608" Oct 03 13:14:54 crc kubenswrapper[4578]: I1003 13:14:54.817589 4578 generic.go:334] "Generic (PLEG): container finished" podID="241fb208-e394-48e3-acf7-5ec5c90ee78a" containerID="03725b4989bcd78d56fe85a2a5a02fb423b38d935130808036ccd0327ba40e2e" exitCode=143 Oct 03 13:14:54 crc kubenswrapper[4578]: I1003 13:14:54.818819 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"241fb208-e394-48e3-acf7-5ec5c90ee78a","Type":"ContainerDied","Data":"03725b4989bcd78d56fe85a2a5a02fb423b38d935130808036ccd0327ba40e2e"} Oct 03 13:14:56 crc kubenswrapper[4578]: I1003 13:14:56.781932 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7b58cf9568-f8nng" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 03 13:14:56 crc kubenswrapper[4578]: I1003 13:14:56.782342 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:14:57 crc kubenswrapper[4578]: I1003 13:14:57.320153 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 13:14:57 crc kubenswrapper[4578]: I1003 13:14:57.320222 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 13:14:57 crc kubenswrapper[4578]: I1003 13:14:57.490168 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 03 13:14:57 crc kubenswrapper[4578]: I1003 13:14:57.490201 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 03 13:14:57 crc kubenswrapper[4578]: I1003 13:14:57.521234 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 03 13:14:57 crc kubenswrapper[4578]: I1003 13:14:57.819142 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 13:14:57 crc kubenswrapper[4578]: I1003 13:14:57.819388 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 13:14:57 crc kubenswrapper[4578]: I1003 13:14:57.850296 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:14:57 crc kubenswrapper[4578]: I1003 13:14:57.861355 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:14:57 crc kubenswrapper[4578]: I1003 13:14:57.886416 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 03 13:14:57 crc kubenswrapper[4578]: I1003 13:14:57.952084 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-45rx4"] Oct 03 13:14:57 crc kubenswrapper[4578]: I1003 13:14:57.957062 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-45rx4" podUID="7e0fa932-47ae-4ed6-94b3-0b2829dbab5f" containerName="dnsmasq-dns" containerID="cri-o://40ae7f33bb143edc84844d6355501b2b08b9e2e05d72567ae4a58998411b19b5" gracePeriod=10 Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.403914 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.404363 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.801879 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.836501 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-ovsdbserver-nb\") pod \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.836601 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-config\") pod \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.836666 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-dns-swift-storage-0\") pod \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.836749 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-dns-svc\") pod \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.836777 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-ovsdbserver-sb\") pod \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.836852 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwgfv\" (UniqueName: \"kubernetes.io/projected/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-kube-api-access-pwgfv\") pod \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\" (UID: \"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f\") " Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.891594 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-kube-api-access-pwgfv" (OuterVolumeSpecName: "kube-api-access-pwgfv") pod "7e0fa932-47ae-4ed6-94b3-0b2829dbab5f" (UID: "7e0fa932-47ae-4ed6-94b3-0b2829dbab5f"). InnerVolumeSpecName "kube-api-access-pwgfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.935840 4578 generic.go:334] "Generic (PLEG): container finished" podID="7e0fa932-47ae-4ed6-94b3-0b2829dbab5f" containerID="40ae7f33bb143edc84844d6355501b2b08b9e2e05d72567ae4a58998411b19b5" exitCode=0 Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.936667 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-45rx4" Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.938579 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwgfv\" (UniqueName: \"kubernetes.io/projected/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-kube-api-access-pwgfv\") on node \"crc\" DevicePath \"\"" Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.958179 4578 generic.go:334] "Generic (PLEG): container finished" podID="1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e" containerID="003a0c1928553820972cefea96c5d5af55961caad7bbb301c6b076d814788eec" exitCode=0 Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.959409 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7e0fa932-47ae-4ed6-94b3-0b2829dbab5f" (UID: "7e0fa932-47ae-4ed6-94b3-0b2829dbab5f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.959493 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-45rx4" event={"ID":"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f","Type":"ContainerDied","Data":"40ae7f33bb143edc84844d6355501b2b08b9e2e05d72567ae4a58998411b19b5"} Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.959533 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-45rx4" event={"ID":"7e0fa932-47ae-4ed6-94b3-0b2829dbab5f","Type":"ContainerDied","Data":"521b7ef2049075390765ad2edf95200d4dc5c8e1ab2374715bbf5d59e51114ab"} Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.959545 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-q4j79" event={"ID":"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e","Type":"ContainerDied","Data":"003a0c1928553820972cefea96c5d5af55961caad7bbb301c6b076d814788eec"} Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.959600 4578 scope.go:117] "RemoveContainer" containerID="40ae7f33bb143edc84844d6355501b2b08b9e2e05d72567ae4a58998411b19b5" Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.980540 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-config" (OuterVolumeSpecName: "config") pod "7e0fa932-47ae-4ed6-94b3-0b2829dbab5f" (UID: "7e0fa932-47ae-4ed6-94b3-0b2829dbab5f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:14:58 crc kubenswrapper[4578]: I1003 13:14:58.984917 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7e0fa932-47ae-4ed6-94b3-0b2829dbab5f" (UID: "7e0fa932-47ae-4ed6-94b3-0b2829dbab5f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:14:59 crc kubenswrapper[4578]: I1003 13:14:59.001734 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7e0fa932-47ae-4ed6-94b3-0b2829dbab5f" (UID: "7e0fa932-47ae-4ed6-94b3-0b2829dbab5f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:14:59 crc kubenswrapper[4578]: I1003 13:14:59.002078 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7e0fa932-47ae-4ed6-94b3-0b2829dbab5f" (UID: "7e0fa932-47ae-4ed6-94b3-0b2829dbab5f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:14:59 crc kubenswrapper[4578]: I1003 13:14:59.023682 4578 scope.go:117] "RemoveContainer" containerID="1556a725891caf1302c447126133d7a876f00aff2ac0a26aca363a8f02981637" Oct 03 13:14:59 crc kubenswrapper[4578]: I1003 13:14:59.040861 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:14:59 crc kubenswrapper[4578]: I1003 13:14:59.041281 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:14:59 crc kubenswrapper[4578]: I1003 13:14:59.041856 4578 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:14:59 crc kubenswrapper[4578]: I1003 13:14:59.041876 4578 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:14:59 crc kubenswrapper[4578]: I1003 13:14:59.041887 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:14:59 crc kubenswrapper[4578]: I1003 13:14:59.056649 4578 scope.go:117] "RemoveContainer" containerID="40ae7f33bb143edc84844d6355501b2b08b9e2e05d72567ae4a58998411b19b5" Oct 03 13:14:59 crc kubenswrapper[4578]: E1003 13:14:59.057679 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40ae7f33bb143edc84844d6355501b2b08b9e2e05d72567ae4a58998411b19b5\": container with ID starting with 40ae7f33bb143edc84844d6355501b2b08b9e2e05d72567ae4a58998411b19b5 not found: ID does not exist" containerID="40ae7f33bb143edc84844d6355501b2b08b9e2e05d72567ae4a58998411b19b5" Oct 03 13:14:59 crc kubenswrapper[4578]: I1003 13:14:59.057728 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40ae7f33bb143edc84844d6355501b2b08b9e2e05d72567ae4a58998411b19b5"} err="failed to get container status \"40ae7f33bb143edc84844d6355501b2b08b9e2e05d72567ae4a58998411b19b5\": rpc error: code = NotFound desc = could not find container \"40ae7f33bb143edc84844d6355501b2b08b9e2e05d72567ae4a58998411b19b5\": container with ID starting with 40ae7f33bb143edc84844d6355501b2b08b9e2e05d72567ae4a58998411b19b5 not found: ID does not exist" Oct 03 13:14:59 crc kubenswrapper[4578]: I1003 13:14:59.057759 4578 scope.go:117] "RemoveContainer" containerID="1556a725891caf1302c447126133d7a876f00aff2ac0a26aca363a8f02981637" Oct 03 13:14:59 crc kubenswrapper[4578]: E1003 13:14:59.058051 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1556a725891caf1302c447126133d7a876f00aff2ac0a26aca363a8f02981637\": container with ID starting with 1556a725891caf1302c447126133d7a876f00aff2ac0a26aca363a8f02981637 not found: ID does not exist" containerID="1556a725891caf1302c447126133d7a876f00aff2ac0a26aca363a8f02981637" Oct 03 13:14:59 crc kubenswrapper[4578]: I1003 13:14:59.058085 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1556a725891caf1302c447126133d7a876f00aff2ac0a26aca363a8f02981637"} err="failed to get container status \"1556a725891caf1302c447126133d7a876f00aff2ac0a26aca363a8f02981637\": rpc error: code = NotFound desc = could not find container \"1556a725891caf1302c447126133d7a876f00aff2ac0a26aca363a8f02981637\": container with ID starting with 1556a725891caf1302c447126133d7a876f00aff2ac0a26aca363a8f02981637 not found: ID does not exist" Oct 03 13:14:59 crc kubenswrapper[4578]: I1003 13:14:59.274602 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-45rx4"] Oct 03 13:14:59 crc kubenswrapper[4578]: I1003 13:14:59.282599 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-45rx4"] Oct 03 13:14:59 crc kubenswrapper[4578]: I1003 13:14:59.793150 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 03 13:14:59 crc kubenswrapper[4578]: I1003 13:14:59.970562 4578 generic.go:334] "Generic (PLEG): container finished" podID="4304b326-aece-4557-904b-31f2fbab5c1f" containerID="2048be9c4225e7db96e0b1128fcfe04007983ee58ba570f83e79eebe69ed52b5" exitCode=0 Oct 03 13:14:59 crc kubenswrapper[4578]: I1003 13:14:59.970619 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-nvwxn" event={"ID":"4304b326-aece-4557-904b-31f2fbab5c1f","Type":"ContainerDied","Data":"2048be9c4225e7db96e0b1128fcfe04007983ee58ba570f83e79eebe69ed52b5"} Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.168686 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz"] Oct 03 13:15:00 crc kubenswrapper[4578]: E1003 13:15:00.169190 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e0fa932-47ae-4ed6-94b3-0b2829dbab5f" containerName="init" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.169215 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e0fa932-47ae-4ed6-94b3-0b2829dbab5f" containerName="init" Oct 03 13:15:00 crc kubenswrapper[4578]: E1003 13:15:00.169266 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e0fa932-47ae-4ed6-94b3-0b2829dbab5f" containerName="dnsmasq-dns" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.169275 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e0fa932-47ae-4ed6-94b3-0b2829dbab5f" containerName="dnsmasq-dns" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.169489 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e0fa932-47ae-4ed6-94b3-0b2829dbab5f" containerName="dnsmasq-dns" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.170278 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.175143 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.175458 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.190365 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz"] Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.285089 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f002fd1d-bac7-406b-ba17-bd3d68d2052b-config-volume\") pod \"collect-profiles-29324955-rk7qz\" (UID: \"f002fd1d-bac7-406b-ba17-bd3d68d2052b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.285159 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmzp2\" (UniqueName: \"kubernetes.io/projected/f002fd1d-bac7-406b-ba17-bd3d68d2052b-kube-api-access-pmzp2\") pod \"collect-profiles-29324955-rk7qz\" (UID: \"f002fd1d-bac7-406b-ba17-bd3d68d2052b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.285227 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f002fd1d-bac7-406b-ba17-bd3d68d2052b-secret-volume\") pod \"collect-profiles-29324955-rk7qz\" (UID: \"f002fd1d-bac7-406b-ba17-bd3d68d2052b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.370612 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-q4j79" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.387054 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f002fd1d-bac7-406b-ba17-bd3d68d2052b-config-volume\") pod \"collect-profiles-29324955-rk7qz\" (UID: \"f002fd1d-bac7-406b-ba17-bd3d68d2052b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.387116 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmzp2\" (UniqueName: \"kubernetes.io/projected/f002fd1d-bac7-406b-ba17-bd3d68d2052b-kube-api-access-pmzp2\") pod \"collect-profiles-29324955-rk7qz\" (UID: \"f002fd1d-bac7-406b-ba17-bd3d68d2052b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.387151 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f002fd1d-bac7-406b-ba17-bd3d68d2052b-secret-volume\") pod \"collect-profiles-29324955-rk7qz\" (UID: \"f002fd1d-bac7-406b-ba17-bd3d68d2052b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.388500 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f002fd1d-bac7-406b-ba17-bd3d68d2052b-config-volume\") pod \"collect-profiles-29324955-rk7qz\" (UID: \"f002fd1d-bac7-406b-ba17-bd3d68d2052b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.400722 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f002fd1d-bac7-406b-ba17-bd3d68d2052b-secret-volume\") pod \"collect-profiles-29324955-rk7qz\" (UID: \"f002fd1d-bac7-406b-ba17-bd3d68d2052b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.418200 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmzp2\" (UniqueName: \"kubernetes.io/projected/f002fd1d-bac7-406b-ba17-bd3d68d2052b-kube-api-access-pmzp2\") pod \"collect-profiles-29324955-rk7qz\" (UID: \"f002fd1d-bac7-406b-ba17-bd3d68d2052b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.488382 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-combined-ca-bundle\") pod \"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e\" (UID: \"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e\") " Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.488496 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-scripts\") pod \"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e\" (UID: \"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e\") " Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.488578 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrscn\" (UniqueName: \"kubernetes.io/projected/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-kube-api-access-wrscn\") pod \"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e\" (UID: \"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e\") " Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.488807 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-config-data\") pod \"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e\" (UID: \"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e\") " Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.493286 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-kube-api-access-wrscn" (OuterVolumeSpecName: "kube-api-access-wrscn") pod "1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e" (UID: "1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e"). InnerVolumeSpecName "kube-api-access-wrscn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.495828 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-scripts" (OuterVolumeSpecName: "scripts") pod "1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e" (UID: "1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.514711 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e" (UID: "1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.520785 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-config-data" (OuterVolumeSpecName: "config-data") pod "1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e" (UID: "1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.545525 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.590773 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrscn\" (UniqueName: \"kubernetes.io/projected/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-kube-api-access-wrscn\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.590938 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.591027 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.591110 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.918944 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e0fa932-47ae-4ed6-94b3-0b2829dbab5f" path="/var/lib/kubelet/pods/7e0fa932-47ae-4ed6-94b3-0b2829dbab5f/volumes" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.987424 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-q4j79" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.988277 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-q4j79" event={"ID":"1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e","Type":"ContainerDied","Data":"c51c0a4c7bfa9d9f5e7521e5ef3cbb5c80a0da9c759b258c2a536d5b410913bc"} Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.988298 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c51c0a4c7bfa9d9f5e7521e5ef3cbb5c80a0da9c759b258c2a536d5b410913bc" Oct 03 13:15:00 crc kubenswrapper[4578]: I1003 13:15:00.991279 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz"] Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.058719 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 13:15:01 crc kubenswrapper[4578]: E1003 13:15:01.060807 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e" containerName="nova-cell1-conductor-db-sync" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.060844 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e" containerName="nova-cell1-conductor-db-sync" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.061192 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e" containerName="nova-cell1-conductor-db-sync" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.062265 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.069956 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.089257 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.206277 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdbce8e5-76ed-4e0c-9649-5c8c783f0986-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fdbce8e5-76ed-4e0c-9649-5c8c783f0986\") " pod="openstack/nova-cell1-conductor-0" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.206475 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6hx8\" (UniqueName: \"kubernetes.io/projected/fdbce8e5-76ed-4e0c-9649-5c8c783f0986-kube-api-access-m6hx8\") pod \"nova-cell1-conductor-0\" (UID: \"fdbce8e5-76ed-4e0c-9649-5c8c783f0986\") " pod="openstack/nova-cell1-conductor-0" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.206523 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdbce8e5-76ed-4e0c-9649-5c8c783f0986-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fdbce8e5-76ed-4e0c-9649-5c8c783f0986\") " pod="openstack/nova-cell1-conductor-0" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.308752 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdbce8e5-76ed-4e0c-9649-5c8c783f0986-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fdbce8e5-76ed-4e0c-9649-5c8c783f0986\") " pod="openstack/nova-cell1-conductor-0" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.308912 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6hx8\" (UniqueName: \"kubernetes.io/projected/fdbce8e5-76ed-4e0c-9649-5c8c783f0986-kube-api-access-m6hx8\") pod \"nova-cell1-conductor-0\" (UID: \"fdbce8e5-76ed-4e0c-9649-5c8c783f0986\") " pod="openstack/nova-cell1-conductor-0" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.308951 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdbce8e5-76ed-4e0c-9649-5c8c783f0986-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fdbce8e5-76ed-4e0c-9649-5c8c783f0986\") " pod="openstack/nova-cell1-conductor-0" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.317185 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdbce8e5-76ed-4e0c-9649-5c8c783f0986-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fdbce8e5-76ed-4e0c-9649-5c8c783f0986\") " pod="openstack/nova-cell1-conductor-0" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.321904 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdbce8e5-76ed-4e0c-9649-5c8c783f0986-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fdbce8e5-76ed-4e0c-9649-5c8c783f0986\") " pod="openstack/nova-cell1-conductor-0" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.340465 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6hx8\" (UniqueName: \"kubernetes.io/projected/fdbce8e5-76ed-4e0c-9649-5c8c783f0986-kube-api-access-m6hx8\") pod \"nova-cell1-conductor-0\" (UID: \"fdbce8e5-76ed-4e0c-9649-5c8c783f0986\") " pod="openstack/nova-cell1-conductor-0" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.382307 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.495235 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-nvwxn" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.620236 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvd5x\" (UniqueName: \"kubernetes.io/projected/4304b326-aece-4557-904b-31f2fbab5c1f-kube-api-access-fvd5x\") pod \"4304b326-aece-4557-904b-31f2fbab5c1f\" (UID: \"4304b326-aece-4557-904b-31f2fbab5c1f\") " Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.620621 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4304b326-aece-4557-904b-31f2fbab5c1f-config-data\") pod \"4304b326-aece-4557-904b-31f2fbab5c1f\" (UID: \"4304b326-aece-4557-904b-31f2fbab5c1f\") " Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.620674 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4304b326-aece-4557-904b-31f2fbab5c1f-scripts\") pod \"4304b326-aece-4557-904b-31f2fbab5c1f\" (UID: \"4304b326-aece-4557-904b-31f2fbab5c1f\") " Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.620750 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4304b326-aece-4557-904b-31f2fbab5c1f-combined-ca-bundle\") pod \"4304b326-aece-4557-904b-31f2fbab5c1f\" (UID: \"4304b326-aece-4557-904b-31f2fbab5c1f\") " Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.629926 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4304b326-aece-4557-904b-31f2fbab5c1f-kube-api-access-fvd5x" (OuterVolumeSpecName: "kube-api-access-fvd5x") pod "4304b326-aece-4557-904b-31f2fbab5c1f" (UID: "4304b326-aece-4557-904b-31f2fbab5c1f"). InnerVolumeSpecName "kube-api-access-fvd5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.641925 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4304b326-aece-4557-904b-31f2fbab5c1f-scripts" (OuterVolumeSpecName: "scripts") pod "4304b326-aece-4557-904b-31f2fbab5c1f" (UID: "4304b326-aece-4557-904b-31f2fbab5c1f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.662190 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4304b326-aece-4557-904b-31f2fbab5c1f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4304b326-aece-4557-904b-31f2fbab5c1f" (UID: "4304b326-aece-4557-904b-31f2fbab5c1f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.669747 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4304b326-aece-4557-904b-31f2fbab5c1f-config-data" (OuterVolumeSpecName: "config-data") pod "4304b326-aece-4557-904b-31f2fbab5c1f" (UID: "4304b326-aece-4557-904b-31f2fbab5c1f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.722800 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4304b326-aece-4557-904b-31f2fbab5c1f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.723082 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvd5x\" (UniqueName: \"kubernetes.io/projected/4304b326-aece-4557-904b-31f2fbab5c1f-kube-api-access-fvd5x\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.723166 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4304b326-aece-4557-904b-31f2fbab5c1f-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.723223 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4304b326-aece-4557-904b-31f2fbab5c1f-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:01 crc kubenswrapper[4578]: I1003 13:15:01.847255 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 03 13:15:01 crc kubenswrapper[4578]: W1003 13:15:01.850198 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdbce8e5_76ed_4e0c_9649_5c8c783f0986.slice/crio-8d6e62578901e990a44f34c9d4cad9c536059c1a1f9e3e632c535bf3795b8cf9 WatchSource:0}: Error finding container 8d6e62578901e990a44f34c9d4cad9c536059c1a1f9e3e632c535bf3795b8cf9: Status 404 returned error can't find the container with id 8d6e62578901e990a44f34c9d4cad9c536059c1a1f9e3e632c535bf3795b8cf9 Oct 03 13:15:02 crc kubenswrapper[4578]: I1003 13:15:02.015959 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-nvwxn" event={"ID":"4304b326-aece-4557-904b-31f2fbab5c1f","Type":"ContainerDied","Data":"a89a20c7cf57d3138a3b21ef7ed631f359c8a3ac9fd0d96ccc8fd07973d457be"} Oct 03 13:15:02 crc kubenswrapper[4578]: I1003 13:15:02.015995 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a89a20c7cf57d3138a3b21ef7ed631f359c8a3ac9fd0d96ccc8fd07973d457be" Oct 03 13:15:02 crc kubenswrapper[4578]: I1003 13:15:02.016064 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-nvwxn" Oct 03 13:15:02 crc kubenswrapper[4578]: I1003 13:15:02.035277 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"fdbce8e5-76ed-4e0c-9649-5c8c783f0986","Type":"ContainerStarted","Data":"3dff3ea4cbf449bd7b09f95a2a589e4e107116689802fac4c8755cd3c43b20e2"} Oct 03 13:15:02 crc kubenswrapper[4578]: I1003 13:15:02.035536 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"fdbce8e5-76ed-4e0c-9649-5c8c783f0986","Type":"ContainerStarted","Data":"8d6e62578901e990a44f34c9d4cad9c536059c1a1f9e3e632c535bf3795b8cf9"} Oct 03 13:15:02 crc kubenswrapper[4578]: I1003 13:15:02.035565 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 03 13:15:02 crc kubenswrapper[4578]: I1003 13:15:02.040196 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz" event={"ID":"f002fd1d-bac7-406b-ba17-bd3d68d2052b","Type":"ContainerDied","Data":"3629c9811602657c37fee60ede1c360c942f31c728f4b833eaebb5494327cb9a"} Oct 03 13:15:02 crc kubenswrapper[4578]: I1003 13:15:02.040442 4578 generic.go:334] "Generic (PLEG): container finished" podID="f002fd1d-bac7-406b-ba17-bd3d68d2052b" containerID="3629c9811602657c37fee60ede1c360c942f31c728f4b833eaebb5494327cb9a" exitCode=0 Oct 03 13:15:02 crc kubenswrapper[4578]: I1003 13:15:02.040735 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz" event={"ID":"f002fd1d-bac7-406b-ba17-bd3d68d2052b","Type":"ContainerStarted","Data":"cdce748e209564cb2b4e3478dbc6ca95e1c06165ae0bd8cfd0fa55a6b8a155d6"} Oct 03 13:15:02 crc kubenswrapper[4578]: I1003 13:15:02.055620 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=1.055603364 podStartE2EDuration="1.055603364s" podCreationTimestamp="2025-10-03 13:15:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:15:02.049601756 +0000 UTC m=+1437.848073940" watchObservedRunningTime="2025-10-03 13:15:02.055603364 +0000 UTC m=+1437.854075548" Oct 03 13:15:02 crc kubenswrapper[4578]: I1003 13:15:02.171951 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:15:02 crc kubenswrapper[4578]: I1003 13:15:02.172658 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b" containerName="nova-api-api" containerID="cri-o://27198dd8e823529e9fd50a1f477c0d9e19e3f4106df8b6afdf29dbba0c13441c" gracePeriod=30 Oct 03 13:15:02 crc kubenswrapper[4578]: I1003 13:15:02.172620 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b" containerName="nova-api-log" containerID="cri-o://fbd39dc5b7e344c6b998d12c45c0c1cfa3556c8e4d54e1a7a57c01105dd03aba" gracePeriod=30 Oct 03 13:15:02 crc kubenswrapper[4578]: I1003 13:15:02.201623 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:15:02 crc kubenswrapper[4578]: I1003 13:15:02.201961 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="fbaa7077-fdc1-49d9-a05b-213ed1f50138" containerName="nova-scheduler-scheduler" containerID="cri-o://546f2dbd38125228a514aeb67874f62b3a4dc0a03aa867e9f2b355d6007f3abb" gracePeriod=30 Oct 03 13:15:02 crc kubenswrapper[4578]: E1003 13:15:02.494529 4578 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="546f2dbd38125228a514aeb67874f62b3a4dc0a03aa867e9f2b355d6007f3abb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 03 13:15:02 crc kubenswrapper[4578]: E1003 13:15:02.496451 4578 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="546f2dbd38125228a514aeb67874f62b3a4dc0a03aa867e9f2b355d6007f3abb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 03 13:15:02 crc kubenswrapper[4578]: E1003 13:15:02.523062 4578 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="546f2dbd38125228a514aeb67874f62b3a4dc0a03aa867e9f2b355d6007f3abb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 03 13:15:02 crc kubenswrapper[4578]: E1003 13:15:02.523150 4578 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="fbaa7077-fdc1-49d9-a05b-213ed1f50138" containerName="nova-scheduler-scheduler" Oct 03 13:15:02 crc kubenswrapper[4578]: I1003 13:15:02.989020 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.052987 4578 generic.go:334] "Generic (PLEG): container finished" podID="3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b" containerID="fbd39dc5b7e344c6b998d12c45c0c1cfa3556c8e4d54e1a7a57c01105dd03aba" exitCode=143 Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.053063 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b","Type":"ContainerDied","Data":"fbd39dc5b7e344c6b998d12c45c0c1cfa3556c8e4d54e1a7a57c01105dd03aba"} Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.055589 4578 generic.go:334] "Generic (PLEG): container finished" podID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerID="de051c4c2fdbf56aef0e53a639dcd8953e6218bb9bdef3f691e52909ccf616ec" exitCode=137 Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.056498 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b58cf9568-f8nng" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.056720 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b58cf9568-f8nng" event={"ID":"a54aea6a-c051-4832-b764-3ef53a9cc3a1","Type":"ContainerDied","Data":"de051c4c2fdbf56aef0e53a639dcd8953e6218bb9bdef3f691e52909ccf616ec"} Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.056780 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b58cf9568-f8nng" event={"ID":"a54aea6a-c051-4832-b764-3ef53a9cc3a1","Type":"ContainerDied","Data":"e283c28782e5f44345615ef81d7b97769693cfdd7bce448773683ccbd06e5f51"} Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.056805 4578 scope.go:117] "RemoveContainer" containerID="de8222a725afcbc467d7921332c4ab3214da8e901c323972f2737910bf0d32c3" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.180167 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a54aea6a-c051-4832-b764-3ef53a9cc3a1-combined-ca-bundle\") pod \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.180209 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a54aea6a-c051-4832-b764-3ef53a9cc3a1-config-data\") pod \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.180254 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a54aea6a-c051-4832-b764-3ef53a9cc3a1-horizon-tls-certs\") pod \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.180287 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a54aea6a-c051-4832-b764-3ef53a9cc3a1-logs\") pod \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.180307 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a54aea6a-c051-4832-b764-3ef53a9cc3a1-horizon-secret-key\") pod \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.180419 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a54aea6a-c051-4832-b764-3ef53a9cc3a1-scripts\") pod \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.180455 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2q9l\" (UniqueName: \"kubernetes.io/projected/a54aea6a-c051-4832-b764-3ef53a9cc3a1-kube-api-access-q2q9l\") pod \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\" (UID: \"a54aea6a-c051-4832-b764-3ef53a9cc3a1\") " Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.183300 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a54aea6a-c051-4832-b764-3ef53a9cc3a1-logs" (OuterVolumeSpecName: "logs") pod "a54aea6a-c051-4832-b764-3ef53a9cc3a1" (UID: "a54aea6a-c051-4832-b764-3ef53a9cc3a1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.193908 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a54aea6a-c051-4832-b764-3ef53a9cc3a1-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a54aea6a-c051-4832-b764-3ef53a9cc3a1" (UID: "a54aea6a-c051-4832-b764-3ef53a9cc3a1"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.194277 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a54aea6a-c051-4832-b764-3ef53a9cc3a1-kube-api-access-q2q9l" (OuterVolumeSpecName: "kube-api-access-q2q9l") pod "a54aea6a-c051-4832-b764-3ef53a9cc3a1" (UID: "a54aea6a-c051-4832-b764-3ef53a9cc3a1"). InnerVolumeSpecName "kube-api-access-q2q9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.222040 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a54aea6a-c051-4832-b764-3ef53a9cc3a1-scripts" (OuterVolumeSpecName: "scripts") pod "a54aea6a-c051-4832-b764-3ef53a9cc3a1" (UID: "a54aea6a-c051-4832-b764-3ef53a9cc3a1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.230151 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a54aea6a-c051-4832-b764-3ef53a9cc3a1-config-data" (OuterVolumeSpecName: "config-data") pod "a54aea6a-c051-4832-b764-3ef53a9cc3a1" (UID: "a54aea6a-c051-4832-b764-3ef53a9cc3a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.245332 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a54aea6a-c051-4832-b764-3ef53a9cc3a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a54aea6a-c051-4832-b764-3ef53a9cc3a1" (UID: "a54aea6a-c051-4832-b764-3ef53a9cc3a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.286104 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a54aea6a-c051-4832-b764-3ef53a9cc3a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.286461 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a54aea6a-c051-4832-b764-3ef53a9cc3a1-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.286471 4578 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a54aea6a-c051-4832-b764-3ef53a9cc3a1-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.286481 4578 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a54aea6a-c051-4832-b764-3ef53a9cc3a1-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.286490 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a54aea6a-c051-4832-b764-3ef53a9cc3a1-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.286498 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2q9l\" (UniqueName: \"kubernetes.io/projected/a54aea6a-c051-4832-b764-3ef53a9cc3a1-kube-api-access-q2q9l\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.298893 4578 scope.go:117] "RemoveContainer" containerID="de051c4c2fdbf56aef0e53a639dcd8953e6218bb9bdef3f691e52909ccf616ec" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.318173 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a54aea6a-c051-4832-b764-3ef53a9cc3a1-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "a54aea6a-c051-4832-b764-3ef53a9cc3a1" (UID: "a54aea6a-c051-4832-b764-3ef53a9cc3a1"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.393239 4578 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a54aea6a-c051-4832-b764-3ef53a9cc3a1-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.396371 4578 scope.go:117] "RemoveContainer" containerID="de8222a725afcbc467d7921332c4ab3214da8e901c323972f2737910bf0d32c3" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.397228 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7b58cf9568-f8nng"] Oct 03 13:15:03 crc kubenswrapper[4578]: E1003 13:15:03.404756 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de8222a725afcbc467d7921332c4ab3214da8e901c323972f2737910bf0d32c3\": container with ID starting with de8222a725afcbc467d7921332c4ab3214da8e901c323972f2737910bf0d32c3 not found: ID does not exist" containerID="de8222a725afcbc467d7921332c4ab3214da8e901c323972f2737910bf0d32c3" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.404796 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de8222a725afcbc467d7921332c4ab3214da8e901c323972f2737910bf0d32c3"} err="failed to get container status \"de8222a725afcbc467d7921332c4ab3214da8e901c323972f2737910bf0d32c3\": rpc error: code = NotFound desc = could not find container \"de8222a725afcbc467d7921332c4ab3214da8e901c323972f2737910bf0d32c3\": container with ID starting with de8222a725afcbc467d7921332c4ab3214da8e901c323972f2737910bf0d32c3 not found: ID does not exist" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.404827 4578 scope.go:117] "RemoveContainer" containerID="de051c4c2fdbf56aef0e53a639dcd8953e6218bb9bdef3f691e52909ccf616ec" Oct 03 13:15:03 crc kubenswrapper[4578]: E1003 13:15:03.405314 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de051c4c2fdbf56aef0e53a639dcd8953e6218bb9bdef3f691e52909ccf616ec\": container with ID starting with de051c4c2fdbf56aef0e53a639dcd8953e6218bb9bdef3f691e52909ccf616ec not found: ID does not exist" containerID="de051c4c2fdbf56aef0e53a639dcd8953e6218bb9bdef3f691e52909ccf616ec" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.405357 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de051c4c2fdbf56aef0e53a639dcd8953e6218bb9bdef3f691e52909ccf616ec"} err="failed to get container status \"de051c4c2fdbf56aef0e53a639dcd8953e6218bb9bdef3f691e52909ccf616ec\": rpc error: code = NotFound desc = could not find container \"de051c4c2fdbf56aef0e53a639dcd8953e6218bb9bdef3f691e52909ccf616ec\": container with ID starting with de051c4c2fdbf56aef0e53a639dcd8953e6218bb9bdef3f691e52909ccf616ec not found: ID does not exist" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.412522 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7b58cf9568-f8nng"] Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.497339 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.595659 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f002fd1d-bac7-406b-ba17-bd3d68d2052b-config-volume\") pod \"f002fd1d-bac7-406b-ba17-bd3d68d2052b\" (UID: \"f002fd1d-bac7-406b-ba17-bd3d68d2052b\") " Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.595737 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f002fd1d-bac7-406b-ba17-bd3d68d2052b-secret-volume\") pod \"f002fd1d-bac7-406b-ba17-bd3d68d2052b\" (UID: \"f002fd1d-bac7-406b-ba17-bd3d68d2052b\") " Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.595857 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmzp2\" (UniqueName: \"kubernetes.io/projected/f002fd1d-bac7-406b-ba17-bd3d68d2052b-kube-api-access-pmzp2\") pod \"f002fd1d-bac7-406b-ba17-bd3d68d2052b\" (UID: \"f002fd1d-bac7-406b-ba17-bd3d68d2052b\") " Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.596513 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f002fd1d-bac7-406b-ba17-bd3d68d2052b-config-volume" (OuterVolumeSpecName: "config-volume") pod "f002fd1d-bac7-406b-ba17-bd3d68d2052b" (UID: "f002fd1d-bac7-406b-ba17-bd3d68d2052b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.599978 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f002fd1d-bac7-406b-ba17-bd3d68d2052b-kube-api-access-pmzp2" (OuterVolumeSpecName: "kube-api-access-pmzp2") pod "f002fd1d-bac7-406b-ba17-bd3d68d2052b" (UID: "f002fd1d-bac7-406b-ba17-bd3d68d2052b"). InnerVolumeSpecName "kube-api-access-pmzp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.610333 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f002fd1d-bac7-406b-ba17-bd3d68d2052b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f002fd1d-bac7-406b-ba17-bd3d68d2052b" (UID: "f002fd1d-bac7-406b-ba17-bd3d68d2052b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.698513 4578 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f002fd1d-bac7-406b-ba17-bd3d68d2052b-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.698545 4578 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f002fd1d-bac7-406b-ba17-bd3d68d2052b-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:03 crc kubenswrapper[4578]: I1003 13:15:03.698558 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmzp2\" (UniqueName: \"kubernetes.io/projected/f002fd1d-bac7-406b-ba17-bd3d68d2052b-kube-api-access-pmzp2\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:04 crc kubenswrapper[4578]: I1003 13:15:04.067966 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz" Oct 03 13:15:04 crc kubenswrapper[4578]: I1003 13:15:04.068012 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz" event={"ID":"f002fd1d-bac7-406b-ba17-bd3d68d2052b","Type":"ContainerDied","Data":"cdce748e209564cb2b4e3478dbc6ca95e1c06165ae0bd8cfd0fa55a6b8a155d6"} Oct 03 13:15:04 crc kubenswrapper[4578]: I1003 13:15:04.068035 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cdce748e209564cb2b4e3478dbc6ca95e1c06165ae0bd8cfd0fa55a6b8a155d6" Oct 03 13:15:04 crc kubenswrapper[4578]: I1003 13:15:04.416180 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:15:04 crc kubenswrapper[4578]: I1003 13:15:04.416729 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="58664f3d-6768-4b3b-a96a-6c22c2c7a8c4" containerName="kube-state-metrics" containerID="cri-o://48e58c784bfce1386f4309f8d72c8b868d8a7048f75cda2ea774eb2f01a36741" gracePeriod=30 Oct 03 13:15:04 crc kubenswrapper[4578]: I1003 13:15:04.913414 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 13:15:04 crc kubenswrapper[4578]: I1003 13:15:04.938958 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" path="/var/lib/kubelet/pods/a54aea6a-c051-4832-b764-3ef53a9cc3a1/volumes" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.022137 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8hdw\" (UniqueName: \"kubernetes.io/projected/58664f3d-6768-4b3b-a96a-6c22c2c7a8c4-kube-api-access-f8hdw\") pod \"58664f3d-6768-4b3b-a96a-6c22c2c7a8c4\" (UID: \"58664f3d-6768-4b3b-a96a-6c22c2c7a8c4\") " Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.032527 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58664f3d-6768-4b3b-a96a-6c22c2c7a8c4-kube-api-access-f8hdw" (OuterVolumeSpecName: "kube-api-access-f8hdw") pod "58664f3d-6768-4b3b-a96a-6c22c2c7a8c4" (UID: "58664f3d-6768-4b3b-a96a-6c22c2c7a8c4"). InnerVolumeSpecName "kube-api-access-f8hdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.078581 4578 generic.go:334] "Generic (PLEG): container finished" podID="58664f3d-6768-4b3b-a96a-6c22c2c7a8c4" containerID="48e58c784bfce1386f4309f8d72c8b868d8a7048f75cda2ea774eb2f01a36741" exitCode=2 Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.078641 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"58664f3d-6768-4b3b-a96a-6c22c2c7a8c4","Type":"ContainerDied","Data":"48e58c784bfce1386f4309f8d72c8b868d8a7048f75cda2ea774eb2f01a36741"} Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.078772 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.079871 4578 scope.go:117] "RemoveContainer" containerID="48e58c784bfce1386f4309f8d72c8b868d8a7048f75cda2ea774eb2f01a36741" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.079799 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"58664f3d-6768-4b3b-a96a-6c22c2c7a8c4","Type":"ContainerDied","Data":"c885218c52ef95393d306068fb1041f26c7f51dc99c539fe6e65c5ba30864457"} Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.091319 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.091363 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.115907 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.130683 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8hdw\" (UniqueName: \"kubernetes.io/projected/58664f3d-6768-4b3b-a96a-6c22c2c7a8c4-kube-api-access-f8hdw\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.131380 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.136973 4578 scope.go:117] "RemoveContainer" containerID="48e58c784bfce1386f4309f8d72c8b868d8a7048f75cda2ea774eb2f01a36741" Oct 03 13:15:05 crc kubenswrapper[4578]: E1003 13:15:05.137511 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48e58c784bfce1386f4309f8d72c8b868d8a7048f75cda2ea774eb2f01a36741\": container with ID starting with 48e58c784bfce1386f4309f8d72c8b868d8a7048f75cda2ea774eb2f01a36741 not found: ID does not exist" containerID="48e58c784bfce1386f4309f8d72c8b868d8a7048f75cda2ea774eb2f01a36741" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.137571 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48e58c784bfce1386f4309f8d72c8b868d8a7048f75cda2ea774eb2f01a36741"} err="failed to get container status \"48e58c784bfce1386f4309f8d72c8b868d8a7048f75cda2ea774eb2f01a36741\": rpc error: code = NotFound desc = could not find container \"48e58c784bfce1386f4309f8d72c8b868d8a7048f75cda2ea774eb2f01a36741\": container with ID starting with 48e58c784bfce1386f4309f8d72c8b868d8a7048f75cda2ea774eb2f01a36741 not found: ID does not exist" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.145312 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:15:05 crc kubenswrapper[4578]: E1003 13:15:05.145730 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerName="horizon" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.145747 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerName="horizon" Oct 03 13:15:05 crc kubenswrapper[4578]: E1003 13:15:05.145762 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f002fd1d-bac7-406b-ba17-bd3d68d2052b" containerName="collect-profiles" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.145773 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="f002fd1d-bac7-406b-ba17-bd3d68d2052b" containerName="collect-profiles" Oct 03 13:15:05 crc kubenswrapper[4578]: E1003 13:15:05.145797 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerName="horizon" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.145803 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerName="horizon" Oct 03 13:15:05 crc kubenswrapper[4578]: E1003 13:15:05.145812 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58664f3d-6768-4b3b-a96a-6c22c2c7a8c4" containerName="kube-state-metrics" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.145817 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="58664f3d-6768-4b3b-a96a-6c22c2c7a8c4" containerName="kube-state-metrics" Oct 03 13:15:05 crc kubenswrapper[4578]: E1003 13:15:05.145835 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4304b326-aece-4557-904b-31f2fbab5c1f" containerName="nova-manage" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.145842 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="4304b326-aece-4557-904b-31f2fbab5c1f" containerName="nova-manage" Oct 03 13:15:05 crc kubenswrapper[4578]: E1003 13:15:05.145866 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerName="horizon-log" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.145871 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerName="horizon-log" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.146719 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerName="horizon" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.146744 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="4304b326-aece-4557-904b-31f2fbab5c1f" containerName="nova-manage" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.146761 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerName="horizon-log" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.146778 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="f002fd1d-bac7-406b-ba17-bd3d68d2052b" containerName="collect-profiles" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.146792 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="58664f3d-6768-4b3b-a96a-6c22c2c7a8c4" containerName="kube-state-metrics" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.146805 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="a54aea6a-c051-4832-b764-3ef53a9cc3a1" containerName="horizon" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.151495 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.155387 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.155430 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.158577 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.333235 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwrdm\" (UniqueName: \"kubernetes.io/projected/95fd518d-9176-4f37-ba66-e636d9794533-kube-api-access-pwrdm\") pod \"kube-state-metrics-0\" (UID: \"95fd518d-9176-4f37-ba66-e636d9794533\") " pod="openstack/kube-state-metrics-0" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.334211 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/95fd518d-9176-4f37-ba66-e636d9794533-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"95fd518d-9176-4f37-ba66-e636d9794533\") " pod="openstack/kube-state-metrics-0" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.334311 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/95fd518d-9176-4f37-ba66-e636d9794533-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"95fd518d-9176-4f37-ba66-e636d9794533\") " pod="openstack/kube-state-metrics-0" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.334436 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95fd518d-9176-4f37-ba66-e636d9794533-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"95fd518d-9176-4f37-ba66-e636d9794533\") " pod="openstack/kube-state-metrics-0" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.438326 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/95fd518d-9176-4f37-ba66-e636d9794533-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"95fd518d-9176-4f37-ba66-e636d9794533\") " pod="openstack/kube-state-metrics-0" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.438404 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/95fd518d-9176-4f37-ba66-e636d9794533-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"95fd518d-9176-4f37-ba66-e636d9794533\") " pod="openstack/kube-state-metrics-0" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.438477 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95fd518d-9176-4f37-ba66-e636d9794533-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"95fd518d-9176-4f37-ba66-e636d9794533\") " pod="openstack/kube-state-metrics-0" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.438590 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwrdm\" (UniqueName: \"kubernetes.io/projected/95fd518d-9176-4f37-ba66-e636d9794533-kube-api-access-pwrdm\") pod \"kube-state-metrics-0\" (UID: \"95fd518d-9176-4f37-ba66-e636d9794533\") " pod="openstack/kube-state-metrics-0" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.443624 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/95fd518d-9176-4f37-ba66-e636d9794533-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"95fd518d-9176-4f37-ba66-e636d9794533\") " pod="openstack/kube-state-metrics-0" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.444071 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/95fd518d-9176-4f37-ba66-e636d9794533-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"95fd518d-9176-4f37-ba66-e636d9794533\") " pod="openstack/kube-state-metrics-0" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.444886 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95fd518d-9176-4f37-ba66-e636d9794533-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"95fd518d-9176-4f37-ba66-e636d9794533\") " pod="openstack/kube-state-metrics-0" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.456158 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwrdm\" (UniqueName: \"kubernetes.io/projected/95fd518d-9176-4f37-ba66-e636d9794533-kube-api-access-pwrdm\") pod \"kube-state-metrics-0\" (UID: \"95fd518d-9176-4f37-ba66-e636d9794533\") " pod="openstack/kube-state-metrics-0" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.480806 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.907759 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.917466 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.969500 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbaa7077-fdc1-49d9-a05b-213ed1f50138-config-data\") pod \"fbaa7077-fdc1-49d9-a05b-213ed1f50138\" (UID: \"fbaa7077-fdc1-49d9-a05b-213ed1f50138\") " Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.969541 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-config-data\") pod \"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b\" (UID: \"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b\") " Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.969572 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbaa7077-fdc1-49d9-a05b-213ed1f50138-combined-ca-bundle\") pod \"fbaa7077-fdc1-49d9-a05b-213ed1f50138\" (UID: \"fbaa7077-fdc1-49d9-a05b-213ed1f50138\") " Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.969595 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-combined-ca-bundle\") pod \"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b\" (UID: \"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b\") " Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.969618 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbhr4\" (UniqueName: \"kubernetes.io/projected/fbaa7077-fdc1-49d9-a05b-213ed1f50138-kube-api-access-cbhr4\") pod \"fbaa7077-fdc1-49d9-a05b-213ed1f50138\" (UID: \"fbaa7077-fdc1-49d9-a05b-213ed1f50138\") " Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.969686 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bclk\" (UniqueName: \"kubernetes.io/projected/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-kube-api-access-7bclk\") pod \"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b\" (UID: \"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b\") " Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.969742 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-logs\") pod \"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b\" (UID: \"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b\") " Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.976728 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-logs" (OuterVolumeSpecName: "logs") pod "3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b" (UID: "3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.981822 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbaa7077-fdc1-49d9-a05b-213ed1f50138-kube-api-access-cbhr4" (OuterVolumeSpecName: "kube-api-access-cbhr4") pod "fbaa7077-fdc1-49d9-a05b-213ed1f50138" (UID: "fbaa7077-fdc1-49d9-a05b-213ed1f50138"). InnerVolumeSpecName "kube-api-access-cbhr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:15:05 crc kubenswrapper[4578]: I1003 13:15:05.983788 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-kube-api-access-7bclk" (OuterVolumeSpecName: "kube-api-access-7bclk") pod "3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b" (UID: "3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b"). InnerVolumeSpecName "kube-api-access-7bclk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.006924 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b" (UID: "3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.007945 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbaa7077-fdc1-49d9-a05b-213ed1f50138-config-data" (OuterVolumeSpecName: "config-data") pod "fbaa7077-fdc1-49d9-a05b-213ed1f50138" (UID: "fbaa7077-fdc1-49d9-a05b-213ed1f50138"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.010475 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-config-data" (OuterVolumeSpecName: "config-data") pod "3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b" (UID: "3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.011573 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbaa7077-fdc1-49d9-a05b-213ed1f50138-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fbaa7077-fdc1-49d9-a05b-213ed1f50138" (UID: "fbaa7077-fdc1-49d9-a05b-213ed1f50138"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.070277 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.071772 4578 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.071794 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbaa7077-fdc1-49d9-a05b-213ed1f50138-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.071805 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.071814 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbaa7077-fdc1-49d9-a05b-213ed1f50138-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.071825 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.071833 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbhr4\" (UniqueName: \"kubernetes.io/projected/fbaa7077-fdc1-49d9-a05b-213ed1f50138-kube-api-access-cbhr4\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.071842 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bclk\" (UniqueName: \"kubernetes.io/projected/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b-kube-api-access-7bclk\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.090564 4578 generic.go:334] "Generic (PLEG): container finished" podID="fbaa7077-fdc1-49d9-a05b-213ed1f50138" containerID="546f2dbd38125228a514aeb67874f62b3a4dc0a03aa867e9f2b355d6007f3abb" exitCode=0 Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.090678 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.090674 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fbaa7077-fdc1-49d9-a05b-213ed1f50138","Type":"ContainerDied","Data":"546f2dbd38125228a514aeb67874f62b3a4dc0a03aa867e9f2b355d6007f3abb"} Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.092036 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fbaa7077-fdc1-49d9-a05b-213ed1f50138","Type":"ContainerDied","Data":"5b409837939e426cfc20b4bb0891b021475d1e86dd9859d47e76a8a11c4b9ab0"} Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.092071 4578 scope.go:117] "RemoveContainer" containerID="546f2dbd38125228a514aeb67874f62b3a4dc0a03aa867e9f2b355d6007f3abb" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.098664 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"95fd518d-9176-4f37-ba66-e636d9794533","Type":"ContainerStarted","Data":"3c05518e86ad8f199700baf49f610eea6a35d2ebb04239f75d2a145e8dac017c"} Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.105012 4578 generic.go:334] "Generic (PLEG): container finished" podID="3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b" containerID="27198dd8e823529e9fd50a1f477c0d9e19e3f4106df8b6afdf29dbba0c13441c" exitCode=0 Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.105055 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b","Type":"ContainerDied","Data":"27198dd8e823529e9fd50a1f477c0d9e19e3f4106df8b6afdf29dbba0c13441c"} Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.105079 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b","Type":"ContainerDied","Data":"14c0258d47c7e75a1ec6d46f19fd09d746173ffea8ce801bc6c93f498813cd02"} Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.105132 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.137622 4578 scope.go:117] "RemoveContainer" containerID="546f2dbd38125228a514aeb67874f62b3a4dc0a03aa867e9f2b355d6007f3abb" Oct 03 13:15:06 crc kubenswrapper[4578]: E1003 13:15:06.138251 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"546f2dbd38125228a514aeb67874f62b3a4dc0a03aa867e9f2b355d6007f3abb\": container with ID starting with 546f2dbd38125228a514aeb67874f62b3a4dc0a03aa867e9f2b355d6007f3abb not found: ID does not exist" containerID="546f2dbd38125228a514aeb67874f62b3a4dc0a03aa867e9f2b355d6007f3abb" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.138289 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"546f2dbd38125228a514aeb67874f62b3a4dc0a03aa867e9f2b355d6007f3abb"} err="failed to get container status \"546f2dbd38125228a514aeb67874f62b3a4dc0a03aa867e9f2b355d6007f3abb\": rpc error: code = NotFound desc = could not find container \"546f2dbd38125228a514aeb67874f62b3a4dc0a03aa867e9f2b355d6007f3abb\": container with ID starting with 546f2dbd38125228a514aeb67874f62b3a4dc0a03aa867e9f2b355d6007f3abb not found: ID does not exist" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.138310 4578 scope.go:117] "RemoveContainer" containerID="27198dd8e823529e9fd50a1f477c0d9e19e3f4106df8b6afdf29dbba0c13441c" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.140864 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.152168 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.166664 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.177594 4578 scope.go:117] "RemoveContainer" containerID="fbd39dc5b7e344c6b998d12c45c0c1cfa3556c8e4d54e1a7a57c01105dd03aba" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.185697 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:15:06 crc kubenswrapper[4578]: E1003 13:15:06.186125 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbaa7077-fdc1-49d9-a05b-213ed1f50138" containerName="nova-scheduler-scheduler" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.186146 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbaa7077-fdc1-49d9-a05b-213ed1f50138" containerName="nova-scheduler-scheduler" Oct 03 13:15:06 crc kubenswrapper[4578]: E1003 13:15:06.186166 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b" containerName="nova-api-api" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.186175 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b" containerName="nova-api-api" Oct 03 13:15:06 crc kubenswrapper[4578]: E1003 13:15:06.186193 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b" containerName="nova-api-log" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.186199 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b" containerName="nova-api-log" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.186365 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b" containerName="nova-api-api" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.186378 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbaa7077-fdc1-49d9-a05b-213ed1f50138" containerName="nova-scheduler-scheduler" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.186391 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b" containerName="nova-api-log" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.187194 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.204060 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.204259 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.211999 4578 scope.go:117] "RemoveContainer" containerID="27198dd8e823529e9fd50a1f477c0d9e19e3f4106df8b6afdf29dbba0c13441c" Oct 03 13:15:06 crc kubenswrapper[4578]: E1003 13:15:06.212353 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27198dd8e823529e9fd50a1f477c0d9e19e3f4106df8b6afdf29dbba0c13441c\": container with ID starting with 27198dd8e823529e9fd50a1f477c0d9e19e3f4106df8b6afdf29dbba0c13441c not found: ID does not exist" containerID="27198dd8e823529e9fd50a1f477c0d9e19e3f4106df8b6afdf29dbba0c13441c" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.212462 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27198dd8e823529e9fd50a1f477c0d9e19e3f4106df8b6afdf29dbba0c13441c"} err="failed to get container status \"27198dd8e823529e9fd50a1f477c0d9e19e3f4106df8b6afdf29dbba0c13441c\": rpc error: code = NotFound desc = could not find container \"27198dd8e823529e9fd50a1f477c0d9e19e3f4106df8b6afdf29dbba0c13441c\": container with ID starting with 27198dd8e823529e9fd50a1f477c0d9e19e3f4106df8b6afdf29dbba0c13441c not found: ID does not exist" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.212562 4578 scope.go:117] "RemoveContainer" containerID="fbd39dc5b7e344c6b998d12c45c0c1cfa3556c8e4d54e1a7a57c01105dd03aba" Oct 03 13:15:06 crc kubenswrapper[4578]: E1003 13:15:06.214934 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbd39dc5b7e344c6b998d12c45c0c1cfa3556c8e4d54e1a7a57c01105dd03aba\": container with ID starting with fbd39dc5b7e344c6b998d12c45c0c1cfa3556c8e4d54e1a7a57c01105dd03aba not found: ID does not exist" containerID="fbd39dc5b7e344c6b998d12c45c0c1cfa3556c8e4d54e1a7a57c01105dd03aba" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.214964 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbd39dc5b7e344c6b998d12c45c0c1cfa3556c8e4d54e1a7a57c01105dd03aba"} err="failed to get container status \"fbd39dc5b7e344c6b998d12c45c0c1cfa3556c8e4d54e1a7a57c01105dd03aba\": rpc error: code = NotFound desc = could not find container \"fbd39dc5b7e344c6b998d12c45c0c1cfa3556c8e4d54e1a7a57c01105dd03aba\": container with ID starting with fbd39dc5b7e344c6b998d12c45c0c1cfa3556c8e4d54e1a7a57c01105dd03aba not found: ID does not exist" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.214948 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.225639 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.227540 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.229919 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.237591 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.277389 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-logs\") pod \"nova-api-0\" (UID: \"0edcba27-85bd-45cd-9eb9-2759fa4b07ca\") " pod="openstack/nova-api-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.277444 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9de81a4c-5730-443b-9a48-1923ddb3c139-config-data\") pod \"nova-scheduler-0\" (UID: \"9de81a4c-5730-443b-9a48-1923ddb3c139\") " pod="openstack/nova-scheduler-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.277490 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0edcba27-85bd-45cd-9eb9-2759fa4b07ca\") " pod="openstack/nova-api-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.277533 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5qt6\" (UniqueName: \"kubernetes.io/projected/9de81a4c-5730-443b-9a48-1923ddb3c139-kube-api-access-r5qt6\") pod \"nova-scheduler-0\" (UID: \"9de81a4c-5730-443b-9a48-1923ddb3c139\") " pod="openstack/nova-scheduler-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.277577 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-config-data\") pod \"nova-api-0\" (UID: \"0edcba27-85bd-45cd-9eb9-2759fa4b07ca\") " pod="openstack/nova-api-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.277618 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9de81a4c-5730-443b-9a48-1923ddb3c139-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9de81a4c-5730-443b-9a48-1923ddb3c139\") " pod="openstack/nova-scheduler-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.277731 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfwpm\" (UniqueName: \"kubernetes.io/projected/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-kube-api-access-xfwpm\") pod \"nova-api-0\" (UID: \"0edcba27-85bd-45cd-9eb9-2759fa4b07ca\") " pod="openstack/nova-api-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.379759 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-config-data\") pod \"nova-api-0\" (UID: \"0edcba27-85bd-45cd-9eb9-2759fa4b07ca\") " pod="openstack/nova-api-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.379826 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9de81a4c-5730-443b-9a48-1923ddb3c139-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9de81a4c-5730-443b-9a48-1923ddb3c139\") " pod="openstack/nova-scheduler-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.379892 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfwpm\" (UniqueName: \"kubernetes.io/projected/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-kube-api-access-xfwpm\") pod \"nova-api-0\" (UID: \"0edcba27-85bd-45cd-9eb9-2759fa4b07ca\") " pod="openstack/nova-api-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.379941 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-logs\") pod \"nova-api-0\" (UID: \"0edcba27-85bd-45cd-9eb9-2759fa4b07ca\") " pod="openstack/nova-api-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.379977 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9de81a4c-5730-443b-9a48-1923ddb3c139-config-data\") pod \"nova-scheduler-0\" (UID: \"9de81a4c-5730-443b-9a48-1923ddb3c139\") " pod="openstack/nova-scheduler-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.380029 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0edcba27-85bd-45cd-9eb9-2759fa4b07ca\") " pod="openstack/nova-api-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.380054 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5qt6\" (UniqueName: \"kubernetes.io/projected/9de81a4c-5730-443b-9a48-1923ddb3c139-kube-api-access-r5qt6\") pod \"nova-scheduler-0\" (UID: \"9de81a4c-5730-443b-9a48-1923ddb3c139\") " pod="openstack/nova-scheduler-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.381542 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-logs\") pod \"nova-api-0\" (UID: \"0edcba27-85bd-45cd-9eb9-2759fa4b07ca\") " pod="openstack/nova-api-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.384945 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-config-data\") pod \"nova-api-0\" (UID: \"0edcba27-85bd-45cd-9eb9-2759fa4b07ca\") " pod="openstack/nova-api-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.385996 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9de81a4c-5730-443b-9a48-1923ddb3c139-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"9de81a4c-5730-443b-9a48-1923ddb3c139\") " pod="openstack/nova-scheduler-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.389558 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9de81a4c-5730-443b-9a48-1923ddb3c139-config-data\") pod \"nova-scheduler-0\" (UID: \"9de81a4c-5730-443b-9a48-1923ddb3c139\") " pod="openstack/nova-scheduler-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.392280 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0edcba27-85bd-45cd-9eb9-2759fa4b07ca\") " pod="openstack/nova-api-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.404533 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfwpm\" (UniqueName: \"kubernetes.io/projected/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-kube-api-access-xfwpm\") pod \"nova-api-0\" (UID: \"0edcba27-85bd-45cd-9eb9-2759fa4b07ca\") " pod="openstack/nova-api-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.405133 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5qt6\" (UniqueName: \"kubernetes.io/projected/9de81a4c-5730-443b-9a48-1923ddb3c139-kube-api-access-r5qt6\") pod \"nova-scheduler-0\" (UID: \"9de81a4c-5730-443b-9a48-1923ddb3c139\") " pod="openstack/nova-scheduler-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.422745 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.515991 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.545614 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.935010 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b" path="/var/lib/kubelet/pods/3b1f1c2c-d1d8-4b74-a471-320b5a3bad6b/volumes" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.936042 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58664f3d-6768-4b3b-a96a-6c22c2c7a8c4" path="/var/lib/kubelet/pods/58664f3d-6768-4b3b-a96a-6c22c2c7a8c4/volumes" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.940673 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbaa7077-fdc1-49d9-a05b-213ed1f50138" path="/var/lib/kubelet/pods/fbaa7077-fdc1-49d9-a05b-213ed1f50138/volumes" Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.941198 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.941484 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4c345497-9620-482e-abfa-868112c6e1e6" containerName="ceilometer-central-agent" containerID="cri-o://2fa6108292cf4712139d664537bc28e717b7be6deb73829c486d76891ae137f1" gracePeriod=30 Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.941506 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4c345497-9620-482e-abfa-868112c6e1e6" containerName="proxy-httpd" containerID="cri-o://9570a50ac3ed7d4fc6eebdda73ed2bba5de9e90a534b24b6494319a43857cf9b" gracePeriod=30 Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.941587 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4c345497-9620-482e-abfa-868112c6e1e6" containerName="ceilometer-notification-agent" containerID="cri-o://2903e87c0cb090ddcda043a9f1ec22e7ed66c8e3fbf355cc766c6bc247368eae" gracePeriod=30 Oct 03 13:15:06 crc kubenswrapper[4578]: I1003 13:15:06.941626 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4c345497-9620-482e-abfa-868112c6e1e6" containerName="sg-core" containerID="cri-o://1dc12dff3e12a4279cd27e0b3cfcb19a393615a985c7380d38cf793c967c9e1a" gracePeriod=30 Oct 03 13:15:07 crc kubenswrapper[4578]: I1003 13:15:07.099333 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:15:07 crc kubenswrapper[4578]: I1003 13:15:07.113872 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9de81a4c-5730-443b-9a48-1923ddb3c139","Type":"ContainerStarted","Data":"5c4160ee31028df1e84536d44f54f1e4fb054b1da3e286e7e1ff849c1246231c"} Oct 03 13:15:07 crc kubenswrapper[4578]: I1003 13:15:07.115408 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"95fd518d-9176-4f37-ba66-e636d9794533","Type":"ContainerStarted","Data":"ea510510285af4c9755be31c93774ac46f60174146359481844aac5c41b8267b"} Oct 03 13:15:07 crc kubenswrapper[4578]: I1003 13:15:07.116791 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 03 13:15:07 crc kubenswrapper[4578]: I1003 13:15:07.119706 4578 generic.go:334] "Generic (PLEG): container finished" podID="4c345497-9620-482e-abfa-868112c6e1e6" containerID="1dc12dff3e12a4279cd27e0b3cfcb19a393615a985c7380d38cf793c967c9e1a" exitCode=2 Oct 03 13:15:07 crc kubenswrapper[4578]: I1003 13:15:07.119785 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c345497-9620-482e-abfa-868112c6e1e6","Type":"ContainerDied","Data":"1dc12dff3e12a4279cd27e0b3cfcb19a393615a985c7380d38cf793c967c9e1a"} Oct 03 13:15:07 crc kubenswrapper[4578]: I1003 13:15:07.217247 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.839233842 podStartE2EDuration="2.217205615s" podCreationTimestamp="2025-10-03 13:15:05 +0000 UTC" firstStartedPulling="2025-10-03 13:15:06.080175176 +0000 UTC m=+1441.878647360" lastFinishedPulling="2025-10-03 13:15:06.458146949 +0000 UTC m=+1442.256619133" observedRunningTime="2025-10-03 13:15:07.147230507 +0000 UTC m=+1442.945702681" watchObservedRunningTime="2025-10-03 13:15:07.217205615 +0000 UTC m=+1443.015677799" Oct 03 13:15:07 crc kubenswrapper[4578]: I1003 13:15:07.226445 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:15:08 crc kubenswrapper[4578]: I1003 13:15:08.131872 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0edcba27-85bd-45cd-9eb9-2759fa4b07ca","Type":"ContainerStarted","Data":"242d22f7ce31b5ac25ceff2ed17b5b4522e4a5b51777882a8a630cfceda49eaf"} Oct 03 13:15:08 crc kubenswrapper[4578]: I1003 13:15:08.132232 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0edcba27-85bd-45cd-9eb9-2759fa4b07ca","Type":"ContainerStarted","Data":"8c5c130602f259d06dc208f8e931341ec87ddd4840dad5aa40a13a0f03673f83"} Oct 03 13:15:08 crc kubenswrapper[4578]: I1003 13:15:08.132246 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0edcba27-85bd-45cd-9eb9-2759fa4b07ca","Type":"ContainerStarted","Data":"3e3a565610e85710f456133e46b8e4357b45a85c276073665e1a02e035be198a"} Oct 03 13:15:08 crc kubenswrapper[4578]: I1003 13:15:08.137390 4578 generic.go:334] "Generic (PLEG): container finished" podID="4c345497-9620-482e-abfa-868112c6e1e6" containerID="9570a50ac3ed7d4fc6eebdda73ed2bba5de9e90a534b24b6494319a43857cf9b" exitCode=0 Oct 03 13:15:08 crc kubenswrapper[4578]: I1003 13:15:08.137413 4578 generic.go:334] "Generic (PLEG): container finished" podID="4c345497-9620-482e-abfa-868112c6e1e6" containerID="2fa6108292cf4712139d664537bc28e717b7be6deb73829c486d76891ae137f1" exitCode=0 Oct 03 13:15:08 crc kubenswrapper[4578]: I1003 13:15:08.137444 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c345497-9620-482e-abfa-868112c6e1e6","Type":"ContainerDied","Data":"9570a50ac3ed7d4fc6eebdda73ed2bba5de9e90a534b24b6494319a43857cf9b"} Oct 03 13:15:08 crc kubenswrapper[4578]: I1003 13:15:08.137464 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c345497-9620-482e-abfa-868112c6e1e6","Type":"ContainerDied","Data":"2fa6108292cf4712139d664537bc28e717b7be6deb73829c486d76891ae137f1"} Oct 03 13:15:08 crc kubenswrapper[4578]: I1003 13:15:08.139669 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9de81a4c-5730-443b-9a48-1923ddb3c139","Type":"ContainerStarted","Data":"84b5a9682c1a09a7c23497c8aa1c6992b406a591d51024911638716e9cca16a5"} Oct 03 13:15:08 crc kubenswrapper[4578]: I1003 13:15:08.153164 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.153144178 podStartE2EDuration="2.153144178s" podCreationTimestamp="2025-10-03 13:15:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:15:08.147625924 +0000 UTC m=+1443.946098108" watchObservedRunningTime="2025-10-03 13:15:08.153144178 +0000 UTC m=+1443.951616362" Oct 03 13:15:08 crc kubenswrapper[4578]: I1003 13:15:08.167136 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.167117017 podStartE2EDuration="2.167117017s" podCreationTimestamp="2025-10-03 13:15:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:15:08.166955821 +0000 UTC m=+1443.965428005" watchObservedRunningTime="2025-10-03 13:15:08.167117017 +0000 UTC m=+1443.965589201" Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.151906 4578 generic.go:334] "Generic (PLEG): container finished" podID="4c345497-9620-482e-abfa-868112c6e1e6" containerID="2903e87c0cb090ddcda043a9f1ec22e7ed66c8e3fbf355cc766c6bc247368eae" exitCode=0 Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.151974 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c345497-9620-482e-abfa-868112c6e1e6","Type":"ContainerDied","Data":"2903e87c0cb090ddcda043a9f1ec22e7ed66c8e3fbf355cc766c6bc247368eae"} Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.152357 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c345497-9620-482e-abfa-868112c6e1e6","Type":"ContainerDied","Data":"027771361b2a1d950bc6dc3775bf77cd0f477140f9338a9d913dc66aee287485"} Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.152422 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="027771361b2a1d950bc6dc3775bf77cd0f477140f9338a9d913dc66aee287485" Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.225238 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.238827 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c345497-9620-482e-abfa-868112c6e1e6-run-httpd\") pod \"4c345497-9620-482e-abfa-868112c6e1e6\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.238927 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c345497-9620-482e-abfa-868112c6e1e6-log-httpd\") pod \"4c345497-9620-482e-abfa-868112c6e1e6\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.238992 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-combined-ca-bundle\") pod \"4c345497-9620-482e-abfa-868112c6e1e6\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.239036 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s57c5\" (UniqueName: \"kubernetes.io/projected/4c345497-9620-482e-abfa-868112c6e1e6-kube-api-access-s57c5\") pod \"4c345497-9620-482e-abfa-868112c6e1e6\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.239095 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-config-data\") pod \"4c345497-9620-482e-abfa-868112c6e1e6\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.239120 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-scripts\") pod \"4c345497-9620-482e-abfa-868112c6e1e6\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.239149 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-sg-core-conf-yaml\") pod \"4c345497-9620-482e-abfa-868112c6e1e6\" (UID: \"4c345497-9620-482e-abfa-868112c6e1e6\") " Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.240263 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c345497-9620-482e-abfa-868112c6e1e6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4c345497-9620-482e-abfa-868112c6e1e6" (UID: "4c345497-9620-482e-abfa-868112c6e1e6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.241030 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c345497-9620-482e-abfa-868112c6e1e6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4c345497-9620-482e-abfa-868112c6e1e6" (UID: "4c345497-9620-482e-abfa-868112c6e1e6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.254930 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c345497-9620-482e-abfa-868112c6e1e6-kube-api-access-s57c5" (OuterVolumeSpecName: "kube-api-access-s57c5") pod "4c345497-9620-482e-abfa-868112c6e1e6" (UID: "4c345497-9620-482e-abfa-868112c6e1e6"). InnerVolumeSpecName "kube-api-access-s57c5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.256337 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-scripts" (OuterVolumeSpecName: "scripts") pod "4c345497-9620-482e-abfa-868112c6e1e6" (UID: "4c345497-9620-482e-abfa-868112c6e1e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.310826 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4c345497-9620-482e-abfa-868112c6e1e6" (UID: "4c345497-9620-482e-abfa-868112c6e1e6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.340984 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s57c5\" (UniqueName: \"kubernetes.io/projected/4c345497-9620-482e-abfa-868112c6e1e6-kube-api-access-s57c5\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.341017 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.341029 4578 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.341039 4578 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c345497-9620-482e-abfa-868112c6e1e6-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.341050 4578 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c345497-9620-482e-abfa-868112c6e1e6-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.394697 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c345497-9620-482e-abfa-868112c6e1e6" (UID: "4c345497-9620-482e-abfa-868112c6e1e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.427439 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-config-data" (OuterVolumeSpecName: "config-data") pod "4c345497-9620-482e-abfa-868112c6e1e6" (UID: "4c345497-9620-482e-abfa-868112c6e1e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.442960 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:09 crc kubenswrapper[4578]: I1003 13:15:09.443176 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c345497-9620-482e-abfa-868112c6e1e6-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.162030 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.223353 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.235737 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.247677 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:15:10 crc kubenswrapper[4578]: E1003 13:15:10.248327 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c345497-9620-482e-abfa-868112c6e1e6" containerName="ceilometer-central-agent" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.248437 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c345497-9620-482e-abfa-868112c6e1e6" containerName="ceilometer-central-agent" Oct 03 13:15:10 crc kubenswrapper[4578]: E1003 13:15:10.248542 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c345497-9620-482e-abfa-868112c6e1e6" containerName="ceilometer-notification-agent" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.248611 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c345497-9620-482e-abfa-868112c6e1e6" containerName="ceilometer-notification-agent" Oct 03 13:15:10 crc kubenswrapper[4578]: E1003 13:15:10.248723 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c345497-9620-482e-abfa-868112c6e1e6" containerName="proxy-httpd" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.248819 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c345497-9620-482e-abfa-868112c6e1e6" containerName="proxy-httpd" Oct 03 13:15:10 crc kubenswrapper[4578]: E1003 13:15:10.248899 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c345497-9620-482e-abfa-868112c6e1e6" containerName="sg-core" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.248958 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c345497-9620-482e-abfa-868112c6e1e6" containerName="sg-core" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.249194 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c345497-9620-482e-abfa-868112c6e1e6" containerName="ceilometer-central-agent" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.249265 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c345497-9620-482e-abfa-868112c6e1e6" containerName="ceilometer-notification-agent" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.249332 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c345497-9620-482e-abfa-868112c6e1e6" containerName="proxy-httpd" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.249406 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c345497-9620-482e-abfa-868112c6e1e6" containerName="sg-core" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.251173 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.253477 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.253759 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.257409 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.257499 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.367477 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.367553 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-scripts\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.367679 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d8e8cfc9-5761-4caf-9702-a0809a596754-log-httpd\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.367704 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.367747 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d8e8cfc9-5761-4caf-9702-a0809a596754-run-httpd\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.367786 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-config-data\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.367877 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.368238 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qzps\" (UniqueName: \"kubernetes.io/projected/d8e8cfc9-5761-4caf-9702-a0809a596754-kube-api-access-9qzps\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.469572 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-scripts\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.469670 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d8e8cfc9-5761-4caf-9702-a0809a596754-log-httpd\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.469690 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.469733 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d8e8cfc9-5761-4caf-9702-a0809a596754-run-httpd\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.469774 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-config-data\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.469792 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.469820 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qzps\" (UniqueName: \"kubernetes.io/projected/d8e8cfc9-5761-4caf-9702-a0809a596754-kube-api-access-9qzps\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.469840 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.470442 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d8e8cfc9-5761-4caf-9702-a0809a596754-run-httpd\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.470568 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d8e8cfc9-5761-4caf-9702-a0809a596754-log-httpd\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.474439 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.474829 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.482191 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-config-data\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.482495 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-scripts\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.486025 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.490248 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qzps\" (UniqueName: \"kubernetes.io/projected/d8e8cfc9-5761-4caf-9702-a0809a596754-kube-api-access-9qzps\") pod \"ceilometer-0\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.575235 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:15:10 crc kubenswrapper[4578]: I1003 13:15:10.924575 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c345497-9620-482e-abfa-868112c6e1e6" path="/var/lib/kubelet/pods/4c345497-9620-482e-abfa-868112c6e1e6/volumes" Oct 03 13:15:11 crc kubenswrapper[4578]: I1003 13:15:11.007688 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:15:11 crc kubenswrapper[4578]: I1003 13:15:11.172592 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d8e8cfc9-5761-4caf-9702-a0809a596754","Type":"ContainerStarted","Data":"d5576c016acdfd1becb1f21311dbb08674d2ad63ab4ab4bb87170462d679ac60"} Oct 03 13:15:11 crc kubenswrapper[4578]: I1003 13:15:11.517401 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 03 13:15:12 crc kubenswrapper[4578]: I1003 13:15:12.183166 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d8e8cfc9-5761-4caf-9702-a0809a596754","Type":"ContainerStarted","Data":"7a24c55735edc8b55d25b89a091ef7ca3d644e8a1affa5bd1b699876fb8b363e"} Oct 03 13:15:13 crc kubenswrapper[4578]: I1003 13:15:13.194881 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d8e8cfc9-5761-4caf-9702-a0809a596754","Type":"ContainerStarted","Data":"1e2850122657463934810cf8beaa25b0b49f53dd5ba5a3c0c3b1b31b9797a6a1"} Oct 03 13:15:13 crc kubenswrapper[4578]: I1003 13:15:13.195213 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d8e8cfc9-5761-4caf-9702-a0809a596754","Type":"ContainerStarted","Data":"239d72981276c68c743a6f7b7a0e09e38b2b12cec6fa41760cf8b5e459e6f02a"} Oct 03 13:15:15 crc kubenswrapper[4578]: I1003 13:15:15.212315 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d8e8cfc9-5761-4caf-9702-a0809a596754","Type":"ContainerStarted","Data":"a7147cc315d8b21fbd71f93362f1f57c969862885b1f0de7294648a8c5016cb8"} Oct 03 13:15:15 crc kubenswrapper[4578]: I1003 13:15:15.212998 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 13:15:15 crc kubenswrapper[4578]: I1003 13:15:15.236978 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.704178703 podStartE2EDuration="5.236960725s" podCreationTimestamp="2025-10-03 13:15:10 +0000 UTC" firstStartedPulling="2025-10-03 13:15:11.009003194 +0000 UTC m=+1446.807475368" lastFinishedPulling="2025-10-03 13:15:14.541785206 +0000 UTC m=+1450.340257390" observedRunningTime="2025-10-03 13:15:15.235047365 +0000 UTC m=+1451.033519549" watchObservedRunningTime="2025-10-03 13:15:15.236960725 +0000 UTC m=+1451.035432909" Oct 03 13:15:15 crc kubenswrapper[4578]: I1003 13:15:15.491441 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 03 13:15:16 crc kubenswrapper[4578]: I1003 13:15:16.516819 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 03 13:15:16 crc kubenswrapper[4578]: I1003 13:15:16.544323 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 03 13:15:16 crc kubenswrapper[4578]: I1003 13:15:16.549276 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 13:15:16 crc kubenswrapper[4578]: I1003 13:15:16.549346 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 13:15:16 crc kubenswrapper[4578]: I1003 13:15:16.864918 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n52vx"] Oct 03 13:15:16 crc kubenswrapper[4578]: I1003 13:15:16.867898 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n52vx" Oct 03 13:15:16 crc kubenswrapper[4578]: I1003 13:15:16.875342 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n52vx"] Oct 03 13:15:16 crc kubenswrapper[4578]: I1003 13:15:16.886860 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn7xl\" (UniqueName: \"kubernetes.io/projected/0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5-kube-api-access-qn7xl\") pod \"community-operators-n52vx\" (UID: \"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5\") " pod="openshift-marketplace/community-operators-n52vx" Oct 03 13:15:16 crc kubenswrapper[4578]: I1003 13:15:16.886919 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5-utilities\") pod \"community-operators-n52vx\" (UID: \"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5\") " pod="openshift-marketplace/community-operators-n52vx" Oct 03 13:15:16 crc kubenswrapper[4578]: I1003 13:15:16.886981 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5-catalog-content\") pod \"community-operators-n52vx\" (UID: \"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5\") " pod="openshift-marketplace/community-operators-n52vx" Oct 03 13:15:16 crc kubenswrapper[4578]: I1003 13:15:16.988991 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn7xl\" (UniqueName: \"kubernetes.io/projected/0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5-kube-api-access-qn7xl\") pod \"community-operators-n52vx\" (UID: \"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5\") " pod="openshift-marketplace/community-operators-n52vx" Oct 03 13:15:16 crc kubenswrapper[4578]: I1003 13:15:16.989054 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5-utilities\") pod \"community-operators-n52vx\" (UID: \"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5\") " pod="openshift-marketplace/community-operators-n52vx" Oct 03 13:15:16 crc kubenswrapper[4578]: I1003 13:15:16.989097 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5-catalog-content\") pod \"community-operators-n52vx\" (UID: \"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5\") " pod="openshift-marketplace/community-operators-n52vx" Oct 03 13:15:16 crc kubenswrapper[4578]: I1003 13:15:16.991188 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5-utilities\") pod \"community-operators-n52vx\" (UID: \"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5\") " pod="openshift-marketplace/community-operators-n52vx" Oct 03 13:15:16 crc kubenswrapper[4578]: I1003 13:15:16.991403 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5-catalog-content\") pod \"community-operators-n52vx\" (UID: \"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5\") " pod="openshift-marketplace/community-operators-n52vx" Oct 03 13:15:17 crc kubenswrapper[4578]: I1003 13:15:17.011185 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn7xl\" (UniqueName: \"kubernetes.io/projected/0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5-kube-api-access-qn7xl\") pod \"community-operators-n52vx\" (UID: \"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5\") " pod="openshift-marketplace/community-operators-n52vx" Oct 03 13:15:17 crc kubenswrapper[4578]: I1003 13:15:17.186275 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n52vx" Oct 03 13:15:17 crc kubenswrapper[4578]: I1003 13:15:17.386304 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 03 13:15:17 crc kubenswrapper[4578]: I1003 13:15:17.633072 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0edcba27-85bd-45cd-9eb9-2759fa4b07ca" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 13:15:17 crc kubenswrapper[4578]: I1003 13:15:17.633692 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0edcba27-85bd-45cd-9eb9-2759fa4b07ca" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 13:15:17 crc kubenswrapper[4578]: I1003 13:15:17.821572 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n52vx"] Oct 03 13:15:18 crc kubenswrapper[4578]: I1003 13:15:18.238403 4578 generic.go:334] "Generic (PLEG): container finished" podID="0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5" containerID="c9e3ccec20bb8a4d4eb4b682d12a2bd984b35012a2dc2c2f97cdbed39edee63a" exitCode=0 Oct 03 13:15:18 crc kubenswrapper[4578]: I1003 13:15:18.238453 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n52vx" event={"ID":"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5","Type":"ContainerDied","Data":"c9e3ccec20bb8a4d4eb4b682d12a2bd984b35012a2dc2c2f97cdbed39edee63a"} Oct 03 13:15:18 crc kubenswrapper[4578]: I1003 13:15:18.238871 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n52vx" event={"ID":"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5","Type":"ContainerStarted","Data":"0077377ed13c7d27bc0e52a0c26abc22958c7149ff9191899b8b0cdbe9e3482c"} Oct 03 13:15:19 crc kubenswrapper[4578]: I1003 13:15:19.249865 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n52vx" event={"ID":"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5","Type":"ContainerStarted","Data":"e58105bd2a630f9aa5ec1ad69fcd66bec0bec5aa10ad13a81309014b56d280c9"} Oct 03 13:15:21 crc kubenswrapper[4578]: I1003 13:15:21.277077 4578 generic.go:334] "Generic (PLEG): container finished" podID="0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5" containerID="e58105bd2a630f9aa5ec1ad69fcd66bec0bec5aa10ad13a81309014b56d280c9" exitCode=0 Oct 03 13:15:21 crc kubenswrapper[4578]: I1003 13:15:21.277160 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n52vx" event={"ID":"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5","Type":"ContainerDied","Data":"e58105bd2a630f9aa5ec1ad69fcd66bec0bec5aa10ad13a81309014b56d280c9"} Oct 03 13:15:22 crc kubenswrapper[4578]: I1003 13:15:22.289451 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n52vx" event={"ID":"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5","Type":"ContainerStarted","Data":"832b6d3b6d35484729f6a4775af16b4793c353fd00defe20bf1a5464b1df2945"} Oct 03 13:15:22 crc kubenswrapper[4578]: I1003 13:15:22.317461 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n52vx" podStartSLOduration=2.875013584 podStartE2EDuration="6.317441617s" podCreationTimestamp="2025-10-03 13:15:16 +0000 UTC" firstStartedPulling="2025-10-03 13:15:18.240252433 +0000 UTC m=+1454.038724617" lastFinishedPulling="2025-10-03 13:15:21.682680466 +0000 UTC m=+1457.481152650" observedRunningTime="2025-10-03 13:15:22.313216464 +0000 UTC m=+1458.111688658" watchObservedRunningTime="2025-10-03 13:15:22.317441617 +0000 UTC m=+1458.115913811" Oct 03 13:15:23 crc kubenswrapper[4578]: E1003 13:15:23.877980 4578 manager.go:1116] Failed to create existing container: /kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b1f1c2c_d1d8_4b74_a471_320b5a3bad6b.slice/crio-14c0258d47c7e75a1ec6d46f19fd09d746173ffea8ce801bc6c93f498813cd02: Error finding container 14c0258d47c7e75a1ec6d46f19fd09d746173ffea8ce801bc6c93f498813cd02: Status 404 returned error can't find the container with id 14c0258d47c7e75a1ec6d46f19fd09d746173ffea8ce801bc6c93f498813cd02 Oct 03 13:15:24 crc kubenswrapper[4578]: E1003 13:15:24.228418 4578 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c345497_9620_482e_abfa_868112c6e1e6.slice/crio-027771361b2a1d950bc6dc3775bf77cd0f477140f9338a9d913dc66aee287485\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c345497_9620_482e_abfa_868112c6e1e6.slice/crio-2903e87c0cb090ddcda043a9f1ec22e7ed66c8e3fbf355cc766c6bc247368eae.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c345497_9620_482e_abfa_868112c6e1e6.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c345497_9620_482e_abfa_868112c6e1e6.slice/crio-conmon-2fa6108292cf4712139d664537bc28e717b7be6deb73829c486d76891ae137f1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c345497_9620_482e_abfa_868112c6e1e6.slice/crio-2fa6108292cf4712139d664537bc28e717b7be6deb73829c486d76891ae137f1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod241fb208_e394_48e3_acf7_5ec5c90ee78a.slice/crio-conmon-88a63d3cbf44911329d647457f6985e80beb070f3eea72c841db8dca447a6e10.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e97ceb4_ded2_487b_b5f5_2707b7dfd3d1.slice/crio-conmon-0d9425a149c0242a4eff40b79c7f0e23a726b37168550e73f9255f72a9e7c36c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e97ceb4_ded2_487b_b5f5_2707b7dfd3d1.slice/crio-0d9425a149c0242a4eff40b79c7f0e23a726b37168550e73f9255f72a9e7c36c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c345497_9620_482e_abfa_868112c6e1e6.slice/crio-conmon-2903e87c0cb090ddcda043a9f1ec22e7ed66c8e3fbf355cc766c6bc247368eae.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod241fb208_e394_48e3_acf7_5ec5c90ee78a.slice/crio-88a63d3cbf44911329d647457f6985e80beb070f3eea72c841db8dca447a6e10.scope\": RecentStats: unable to find data in memory cache]" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.284948 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.334186 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1-combined-ca-bundle\") pod \"7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1\" (UID: \"7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1\") " Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.334323 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1-config-data\") pod \"7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1\" (UID: \"7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1\") " Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.334399 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdwcg\" (UniqueName: \"kubernetes.io/projected/7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1-kube-api-access-cdwcg\") pod \"7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1\" (UID: \"7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1\") " Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.337198 4578 generic.go:334] "Generic (PLEG): container finished" podID="7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1" containerID="0d9425a149c0242a4eff40b79c7f0e23a726b37168550e73f9255f72a9e7c36c" exitCode=137 Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.337266 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1","Type":"ContainerDied","Data":"0d9425a149c0242a4eff40b79c7f0e23a726b37168550e73f9255f72a9e7c36c"} Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.337292 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1","Type":"ContainerDied","Data":"69c6104376243ce1eb0f873f51086ba110fed41fd227ba0492fa5f270ab9316f"} Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.337308 4578 scope.go:117] "RemoveContainer" containerID="0d9425a149c0242a4eff40b79c7f0e23a726b37168550e73f9255f72a9e7c36c" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.337418 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.341033 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1-kube-api-access-cdwcg" (OuterVolumeSpecName: "kube-api-access-cdwcg") pod "7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1" (UID: "7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1"). InnerVolumeSpecName "kube-api-access-cdwcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.342261 4578 generic.go:334] "Generic (PLEG): container finished" podID="241fb208-e394-48e3-acf7-5ec5c90ee78a" containerID="88a63d3cbf44911329d647457f6985e80beb070f3eea72c841db8dca447a6e10" exitCode=137 Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.342294 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"241fb208-e394-48e3-acf7-5ec5c90ee78a","Type":"ContainerDied","Data":"88a63d3cbf44911329d647457f6985e80beb070f3eea72c841db8dca447a6e10"} Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.342317 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"241fb208-e394-48e3-acf7-5ec5c90ee78a","Type":"ContainerDied","Data":"1f24033bd84a2f7d1f850101a2765a6d3c9cfc6e0f7da4e07db534bdf2e96cce"} Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.342328 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f24033bd84a2f7d1f850101a2765a6d3c9cfc6e0f7da4e07db534bdf2e96cce" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.364771 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1-config-data" (OuterVolumeSpecName: "config-data") pod "7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1" (UID: "7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.368761 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1" (UID: "7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.405176 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.407679 4578 scope.go:117] "RemoveContainer" containerID="0d9425a149c0242a4eff40b79c7f0e23a726b37168550e73f9255f72a9e7c36c" Oct 03 13:15:24 crc kubenswrapper[4578]: E1003 13:15:24.408125 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d9425a149c0242a4eff40b79c7f0e23a726b37168550e73f9255f72a9e7c36c\": container with ID starting with 0d9425a149c0242a4eff40b79c7f0e23a726b37168550e73f9255f72a9e7c36c not found: ID does not exist" containerID="0d9425a149c0242a4eff40b79c7f0e23a726b37168550e73f9255f72a9e7c36c" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.408169 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d9425a149c0242a4eff40b79c7f0e23a726b37168550e73f9255f72a9e7c36c"} err="failed to get container status \"0d9425a149c0242a4eff40b79c7f0e23a726b37168550e73f9255f72a9e7c36c\": rpc error: code = NotFound desc = could not find container \"0d9425a149c0242a4eff40b79c7f0e23a726b37168550e73f9255f72a9e7c36c\": container with ID starting with 0d9425a149c0242a4eff40b79c7f0e23a726b37168550e73f9255f72a9e7c36c not found: ID does not exist" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.436269 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241fb208-e394-48e3-acf7-5ec5c90ee78a-combined-ca-bundle\") pod \"241fb208-e394-48e3-acf7-5ec5c90ee78a\" (UID: \"241fb208-e394-48e3-acf7-5ec5c90ee78a\") " Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.436527 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/241fb208-e394-48e3-acf7-5ec5c90ee78a-logs\") pod \"241fb208-e394-48e3-acf7-5ec5c90ee78a\" (UID: \"241fb208-e394-48e3-acf7-5ec5c90ee78a\") " Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.436590 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/241fb208-e394-48e3-acf7-5ec5c90ee78a-config-data\") pod \"241fb208-e394-48e3-acf7-5ec5c90ee78a\" (UID: \"241fb208-e394-48e3-acf7-5ec5c90ee78a\") " Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.436614 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z767w\" (UniqueName: \"kubernetes.io/projected/241fb208-e394-48e3-acf7-5ec5c90ee78a-kube-api-access-z767w\") pod \"241fb208-e394-48e3-acf7-5ec5c90ee78a\" (UID: \"241fb208-e394-48e3-acf7-5ec5c90ee78a\") " Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.436933 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/241fb208-e394-48e3-acf7-5ec5c90ee78a-logs" (OuterVolumeSpecName: "logs") pod "241fb208-e394-48e3-acf7-5ec5c90ee78a" (UID: "241fb208-e394-48e3-acf7-5ec5c90ee78a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.437109 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.437122 4578 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/241fb208-e394-48e3-acf7-5ec5c90ee78a-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.437131 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.437141 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdwcg\" (UniqueName: \"kubernetes.io/projected/7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1-kube-api-access-cdwcg\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.441907 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/241fb208-e394-48e3-acf7-5ec5c90ee78a-kube-api-access-z767w" (OuterVolumeSpecName: "kube-api-access-z767w") pod "241fb208-e394-48e3-acf7-5ec5c90ee78a" (UID: "241fb208-e394-48e3-acf7-5ec5c90ee78a"). InnerVolumeSpecName "kube-api-access-z767w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.460402 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241fb208-e394-48e3-acf7-5ec5c90ee78a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "241fb208-e394-48e3-acf7-5ec5c90ee78a" (UID: "241fb208-e394-48e3-acf7-5ec5c90ee78a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.468121 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241fb208-e394-48e3-acf7-5ec5c90ee78a-config-data" (OuterVolumeSpecName: "config-data") pod "241fb208-e394-48e3-acf7-5ec5c90ee78a" (UID: "241fb208-e394-48e3-acf7-5ec5c90ee78a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.538704 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241fb208-e394-48e3-acf7-5ec5c90ee78a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.538750 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z767w\" (UniqueName: \"kubernetes.io/projected/241fb208-e394-48e3-acf7-5ec5c90ee78a-kube-api-access-z767w\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.538767 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/241fb208-e394-48e3-acf7-5ec5c90ee78a-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.668619 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.677465 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.691423 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:15:24 crc kubenswrapper[4578]: E1003 13:15:24.692047 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1" containerName="nova-cell1-novncproxy-novncproxy" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.692066 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1" containerName="nova-cell1-novncproxy-novncproxy" Oct 03 13:15:24 crc kubenswrapper[4578]: E1003 13:15:24.692079 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="241fb208-e394-48e3-acf7-5ec5c90ee78a" containerName="nova-metadata-metadata" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.692087 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="241fb208-e394-48e3-acf7-5ec5c90ee78a" containerName="nova-metadata-metadata" Oct 03 13:15:24 crc kubenswrapper[4578]: E1003 13:15:24.692098 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="241fb208-e394-48e3-acf7-5ec5c90ee78a" containerName="nova-metadata-log" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.692107 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="241fb208-e394-48e3-acf7-5ec5c90ee78a" containerName="nova-metadata-log" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.692319 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="241fb208-e394-48e3-acf7-5ec5c90ee78a" containerName="nova-metadata-metadata" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.692334 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1" containerName="nova-cell1-novncproxy-novncproxy" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.692340 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="241fb208-e394-48e3-acf7-5ec5c90ee78a" containerName="nova-metadata-log" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.692931 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.694723 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.698747 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.698778 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.712305 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.740950 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/624416e7-f2e2-462e-a0f1-e06ce21514f3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"624416e7-f2e2-462e-a0f1-e06ce21514f3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.741012 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/624416e7-f2e2-462e-a0f1-e06ce21514f3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"624416e7-f2e2-462e-a0f1-e06ce21514f3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.741052 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hbx7\" (UniqueName: \"kubernetes.io/projected/624416e7-f2e2-462e-a0f1-e06ce21514f3-kube-api-access-5hbx7\") pod \"nova-cell1-novncproxy-0\" (UID: \"624416e7-f2e2-462e-a0f1-e06ce21514f3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.741073 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/624416e7-f2e2-462e-a0f1-e06ce21514f3-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"624416e7-f2e2-462e-a0f1-e06ce21514f3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.741148 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/624416e7-f2e2-462e-a0f1-e06ce21514f3-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"624416e7-f2e2-462e-a0f1-e06ce21514f3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.842782 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/624416e7-f2e2-462e-a0f1-e06ce21514f3-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"624416e7-f2e2-462e-a0f1-e06ce21514f3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.842860 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/624416e7-f2e2-462e-a0f1-e06ce21514f3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"624416e7-f2e2-462e-a0f1-e06ce21514f3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.842918 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/624416e7-f2e2-462e-a0f1-e06ce21514f3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"624416e7-f2e2-462e-a0f1-e06ce21514f3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.842965 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hbx7\" (UniqueName: \"kubernetes.io/projected/624416e7-f2e2-462e-a0f1-e06ce21514f3-kube-api-access-5hbx7\") pod \"nova-cell1-novncproxy-0\" (UID: \"624416e7-f2e2-462e-a0f1-e06ce21514f3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.842996 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/624416e7-f2e2-462e-a0f1-e06ce21514f3-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"624416e7-f2e2-462e-a0f1-e06ce21514f3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.847069 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/624416e7-f2e2-462e-a0f1-e06ce21514f3-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"624416e7-f2e2-462e-a0f1-e06ce21514f3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.847092 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/624416e7-f2e2-462e-a0f1-e06ce21514f3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"624416e7-f2e2-462e-a0f1-e06ce21514f3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.848439 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/624416e7-f2e2-462e-a0f1-e06ce21514f3-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"624416e7-f2e2-462e-a0f1-e06ce21514f3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.853066 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/624416e7-f2e2-462e-a0f1-e06ce21514f3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"624416e7-f2e2-462e-a0f1-e06ce21514f3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.859204 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hbx7\" (UniqueName: \"kubernetes.io/projected/624416e7-f2e2-462e-a0f1-e06ce21514f3-kube-api-access-5hbx7\") pod \"nova-cell1-novncproxy-0\" (UID: \"624416e7-f2e2-462e-a0f1-e06ce21514f3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:24 crc kubenswrapper[4578]: I1003 13:15:24.930837 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1" path="/var/lib/kubelet/pods/7e97ceb4-ded2-487b-b5f5-2707b7dfd3d1/volumes" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.013372 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.352365 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.379883 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.387377 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.408769 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.417428 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.421964 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.424465 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.430482 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.455390 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cldsw\" (UniqueName: \"kubernetes.io/projected/34b83c0e-8c44-4ceb-abae-35551eb3f183-kube-api-access-cldsw\") pod \"nova-metadata-0\" (UID: \"34b83c0e-8c44-4ceb-abae-35551eb3f183\") " pod="openstack/nova-metadata-0" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.455723 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34b83c0e-8c44-4ceb-abae-35551eb3f183-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"34b83c0e-8c44-4ceb-abae-35551eb3f183\") " pod="openstack/nova-metadata-0" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.455921 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/34b83c0e-8c44-4ceb-abae-35551eb3f183-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"34b83c0e-8c44-4ceb-abae-35551eb3f183\") " pod="openstack/nova-metadata-0" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.456038 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34b83c0e-8c44-4ceb-abae-35551eb3f183-logs\") pod \"nova-metadata-0\" (UID: \"34b83c0e-8c44-4ceb-abae-35551eb3f183\") " pod="openstack/nova-metadata-0" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.456130 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34b83c0e-8c44-4ceb-abae-35551eb3f183-config-data\") pod \"nova-metadata-0\" (UID: \"34b83c0e-8c44-4ceb-abae-35551eb3f183\") " pod="openstack/nova-metadata-0" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.504047 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 03 13:15:25 crc kubenswrapper[4578]: W1003 13:15:25.509432 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod624416e7_f2e2_462e_a0f1_e06ce21514f3.slice/crio-70457b1fbf01dd658c26e73159eda629fcff167277647f83e6415e3c7aae5c4d WatchSource:0}: Error finding container 70457b1fbf01dd658c26e73159eda629fcff167277647f83e6415e3c7aae5c4d: Status 404 returned error can't find the container with id 70457b1fbf01dd658c26e73159eda629fcff167277647f83e6415e3c7aae5c4d Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.558103 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34b83c0e-8c44-4ceb-abae-35551eb3f183-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"34b83c0e-8c44-4ceb-abae-35551eb3f183\") " pod="openstack/nova-metadata-0" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.558638 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/34b83c0e-8c44-4ceb-abae-35551eb3f183-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"34b83c0e-8c44-4ceb-abae-35551eb3f183\") " pod="openstack/nova-metadata-0" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.558671 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34b83c0e-8c44-4ceb-abae-35551eb3f183-logs\") pod \"nova-metadata-0\" (UID: \"34b83c0e-8c44-4ceb-abae-35551eb3f183\") " pod="openstack/nova-metadata-0" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.558689 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34b83c0e-8c44-4ceb-abae-35551eb3f183-config-data\") pod \"nova-metadata-0\" (UID: \"34b83c0e-8c44-4ceb-abae-35551eb3f183\") " pod="openstack/nova-metadata-0" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.558766 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cldsw\" (UniqueName: \"kubernetes.io/projected/34b83c0e-8c44-4ceb-abae-35551eb3f183-kube-api-access-cldsw\") pod \"nova-metadata-0\" (UID: \"34b83c0e-8c44-4ceb-abae-35551eb3f183\") " pod="openstack/nova-metadata-0" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.559238 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34b83c0e-8c44-4ceb-abae-35551eb3f183-logs\") pod \"nova-metadata-0\" (UID: \"34b83c0e-8c44-4ceb-abae-35551eb3f183\") " pod="openstack/nova-metadata-0" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.567619 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34b83c0e-8c44-4ceb-abae-35551eb3f183-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"34b83c0e-8c44-4ceb-abae-35551eb3f183\") " pod="openstack/nova-metadata-0" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.569126 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/34b83c0e-8c44-4ceb-abae-35551eb3f183-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"34b83c0e-8c44-4ceb-abae-35551eb3f183\") " pod="openstack/nova-metadata-0" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.572138 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34b83c0e-8c44-4ceb-abae-35551eb3f183-config-data\") pod \"nova-metadata-0\" (UID: \"34b83c0e-8c44-4ceb-abae-35551eb3f183\") " pod="openstack/nova-metadata-0" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.576112 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cldsw\" (UniqueName: \"kubernetes.io/projected/34b83c0e-8c44-4ceb-abae-35551eb3f183-kube-api-access-cldsw\") pod \"nova-metadata-0\" (UID: \"34b83c0e-8c44-4ceb-abae-35551eb3f183\") " pod="openstack/nova-metadata-0" Oct 03 13:15:25 crc kubenswrapper[4578]: I1003 13:15:25.740821 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:15:26 crc kubenswrapper[4578]: I1003 13:15:26.200691 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:15:26 crc kubenswrapper[4578]: I1003 13:15:26.374523 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"34b83c0e-8c44-4ceb-abae-35551eb3f183","Type":"ContainerStarted","Data":"da4a0e9f1b027c0e7415a9c4d500f0eaddc681a58f73c0316c17f5ad0cf9b7f8"} Oct 03 13:15:26 crc kubenswrapper[4578]: I1003 13:15:26.376824 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"624416e7-f2e2-462e-a0f1-e06ce21514f3","Type":"ContainerStarted","Data":"863fe4c2069b0b50299d56a1699a9244d1c28ab23ab4c2bb1f3a988aa2956d42"} Oct 03 13:15:26 crc kubenswrapper[4578]: I1003 13:15:26.376869 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"624416e7-f2e2-462e-a0f1-e06ce21514f3","Type":"ContainerStarted","Data":"70457b1fbf01dd658c26e73159eda629fcff167277647f83e6415e3c7aae5c4d"} Oct 03 13:15:26 crc kubenswrapper[4578]: I1003 13:15:26.403854 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.4038369 podStartE2EDuration="2.4038369s" podCreationTimestamp="2025-10-03 13:15:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:15:26.399800243 +0000 UTC m=+1462.198272427" watchObservedRunningTime="2025-10-03 13:15:26.4038369 +0000 UTC m=+1462.202309084" Oct 03 13:15:26 crc kubenswrapper[4578]: I1003 13:15:26.563615 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 03 13:15:26 crc kubenswrapper[4578]: I1003 13:15:26.565058 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 03 13:15:26 crc kubenswrapper[4578]: I1003 13:15:26.567593 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 03 13:15:26 crc kubenswrapper[4578]: I1003 13:15:26.571981 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 03 13:15:26 crc kubenswrapper[4578]: I1003 13:15:26.919764 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="241fb208-e394-48e3-acf7-5ec5c90ee78a" path="/var/lib/kubelet/pods/241fb208-e394-48e3-acf7-5ec5c90ee78a/volumes" Oct 03 13:15:27 crc kubenswrapper[4578]: I1003 13:15:27.190060 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n52vx" Oct 03 13:15:27 crc kubenswrapper[4578]: I1003 13:15:27.190473 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n52vx" Oct 03 13:15:27 crc kubenswrapper[4578]: I1003 13:15:27.248707 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n52vx" Oct 03 13:15:27 crc kubenswrapper[4578]: I1003 13:15:27.387116 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"34b83c0e-8c44-4ceb-abae-35551eb3f183","Type":"ContainerStarted","Data":"e1cbf44efb40d92f0202b86942cdf00f47b5fbd2338ecf5dcd010d207d07a59f"} Oct 03 13:15:27 crc kubenswrapper[4578]: I1003 13:15:27.387166 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"34b83c0e-8c44-4ceb-abae-35551eb3f183","Type":"ContainerStarted","Data":"4b057177e19710c3252908b4c28cc2e673d957c473c87d7dc0e2c3c5305d9647"} Oct 03 13:15:27 crc kubenswrapper[4578]: I1003 13:15:27.388245 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 03 13:15:27 crc kubenswrapper[4578]: I1003 13:15:27.393076 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 03 13:15:27 crc kubenswrapper[4578]: I1003 13:15:27.414749 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.414732458 podStartE2EDuration="2.414732458s" podCreationTimestamp="2025-10-03 13:15:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:15:27.408309686 +0000 UTC m=+1463.206781880" watchObservedRunningTime="2025-10-03 13:15:27.414732458 +0000 UTC m=+1463.213204642" Oct 03 13:15:27 crc kubenswrapper[4578]: I1003 13:15:27.448496 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n52vx" Oct 03 13:15:27 crc kubenswrapper[4578]: I1003 13:15:27.548695 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n52vx"] Oct 03 13:15:27 crc kubenswrapper[4578]: I1003 13:15:27.646729 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-wrzpd"] Oct 03 13:15:27 crc kubenswrapper[4578]: I1003 13:15:27.648443 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:27.665496 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-wrzpd"] Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:27.797877 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-wrzpd\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:27.797945 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-wrzpd\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:27.798259 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-wrzpd\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:27.798386 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-wrzpd\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:27.798416 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzrjc\" (UniqueName: \"kubernetes.io/projected/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-kube-api-access-jzrjc\") pod \"dnsmasq-dns-59cf4bdb65-wrzpd\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:27.798460 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-config\") pod \"dnsmasq-dns-59cf4bdb65-wrzpd\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:27.900079 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-wrzpd\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:27.900120 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzrjc\" (UniqueName: \"kubernetes.io/projected/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-kube-api-access-jzrjc\") pod \"dnsmasq-dns-59cf4bdb65-wrzpd\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:27.900155 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-config\") pod \"dnsmasq-dns-59cf4bdb65-wrzpd\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:27.900182 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-wrzpd\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:27.900214 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-wrzpd\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:27.900315 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-wrzpd\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:27.901152 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-wrzpd\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:27.901313 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-wrzpd\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:27.901893 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-wrzpd\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:27.902055 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-wrzpd\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:27.903373 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-config\") pod \"dnsmasq-dns-59cf4bdb65-wrzpd\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:27.924166 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzrjc\" (UniqueName: \"kubernetes.io/projected/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-kube-api-access-jzrjc\") pod \"dnsmasq-dns-59cf4bdb65-wrzpd\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:27.991303 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:28 crc kubenswrapper[4578]: I1003 13:15:28.662099 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-wrzpd"] Oct 03 13:15:28 crc kubenswrapper[4578]: W1003 13:15:28.676344 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6fe75e3a_2d2b_4b6a_a7f4_521fe5f456df.slice/crio-8b545aac0e5015a402f4017420104302b2a9f29e984ad903ea3f04f22ddc2803 WatchSource:0}: Error finding container 8b545aac0e5015a402f4017420104302b2a9f29e984ad903ea3f04f22ddc2803: Status 404 returned error can't find the container with id 8b545aac0e5015a402f4017420104302b2a9f29e984ad903ea3f04f22ddc2803 Oct 03 13:15:29 crc kubenswrapper[4578]: I1003 13:15:29.408804 4578 generic.go:334] "Generic (PLEG): container finished" podID="6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df" containerID="6fa9e91c48b7d8258eabf25b71cd390230210d941e72e5a6c52307399d53284c" exitCode=0 Oct 03 13:15:29 crc kubenswrapper[4578]: I1003 13:15:29.409537 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n52vx" podUID="0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5" containerName="registry-server" containerID="cri-o://832b6d3b6d35484729f6a4775af16b4793c353fd00defe20bf1a5464b1df2945" gracePeriod=2 Oct 03 13:15:29 crc kubenswrapper[4578]: I1003 13:15:29.409181 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" event={"ID":"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df","Type":"ContainerDied","Data":"6fa9e91c48b7d8258eabf25b71cd390230210d941e72e5a6c52307399d53284c"} Oct 03 13:15:29 crc kubenswrapper[4578]: I1003 13:15:29.409930 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" event={"ID":"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df","Type":"ContainerStarted","Data":"8b545aac0e5015a402f4017420104302b2a9f29e984ad903ea3f04f22ddc2803"} Oct 03 13:15:29 crc kubenswrapper[4578]: I1003 13:15:29.912436 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n52vx" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.014271 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.049232 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5-catalog-content\") pod \"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5\" (UID: \"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5\") " Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.049328 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qn7xl\" (UniqueName: \"kubernetes.io/projected/0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5-kube-api-access-qn7xl\") pod \"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5\" (UID: \"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5\") " Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.049390 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5-utilities\") pod \"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5\" (UID: \"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5\") " Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.050710 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5-utilities" (OuterVolumeSpecName: "utilities") pod "0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5" (UID: "0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.070549 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5-kube-api-access-qn7xl" (OuterVolumeSpecName: "kube-api-access-qn7xl") pod "0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5" (UID: "0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5"). InnerVolumeSpecName "kube-api-access-qn7xl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.142787 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5" (UID: "0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.153048 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.153080 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qn7xl\" (UniqueName: \"kubernetes.io/projected/0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5-kube-api-access-qn7xl\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.153090 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.420857 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" event={"ID":"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df","Type":"ContainerStarted","Data":"a930896ada2bd8eb0a5a065becdd4f556524d6303bd4f82d75e6654137c045d7"} Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.421078 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.423231 4578 generic.go:334] "Generic (PLEG): container finished" podID="0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5" containerID="832b6d3b6d35484729f6a4775af16b4793c353fd00defe20bf1a5464b1df2945" exitCode=0 Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.423259 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n52vx" event={"ID":"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5","Type":"ContainerDied","Data":"832b6d3b6d35484729f6a4775af16b4793c353fd00defe20bf1a5464b1df2945"} Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.423287 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n52vx" event={"ID":"0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5","Type":"ContainerDied","Data":"0077377ed13c7d27bc0e52a0c26abc22958c7149ff9191899b8b0cdbe9e3482c"} Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.423306 4578 scope.go:117] "RemoveContainer" containerID="832b6d3b6d35484729f6a4775af16b4793c353fd00defe20bf1a5464b1df2945" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.423332 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n52vx" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.448837 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" podStartSLOduration=3.448821203 podStartE2EDuration="3.448821203s" podCreationTimestamp="2025-10-03 13:15:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:15:30.445238531 +0000 UTC m=+1466.243710735" watchObservedRunningTime="2025-10-03 13:15:30.448821203 +0000 UTC m=+1466.247293387" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.456552 4578 scope.go:117] "RemoveContainer" containerID="e58105bd2a630f9aa5ec1ad69fcd66bec0bec5aa10ad13a81309014b56d280c9" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.486216 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n52vx"] Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.490170 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n52vx"] Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.492262 4578 scope.go:117] "RemoveContainer" containerID="c9e3ccec20bb8a4d4eb4b682d12a2bd984b35012a2dc2c2f97cdbed39edee63a" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.521544 4578 scope.go:117] "RemoveContainer" containerID="832b6d3b6d35484729f6a4775af16b4793c353fd00defe20bf1a5464b1df2945" Oct 03 13:15:30 crc kubenswrapper[4578]: E1003 13:15:30.522062 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"832b6d3b6d35484729f6a4775af16b4793c353fd00defe20bf1a5464b1df2945\": container with ID starting with 832b6d3b6d35484729f6a4775af16b4793c353fd00defe20bf1a5464b1df2945 not found: ID does not exist" containerID="832b6d3b6d35484729f6a4775af16b4793c353fd00defe20bf1a5464b1df2945" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.522099 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"832b6d3b6d35484729f6a4775af16b4793c353fd00defe20bf1a5464b1df2945"} err="failed to get container status \"832b6d3b6d35484729f6a4775af16b4793c353fd00defe20bf1a5464b1df2945\": rpc error: code = NotFound desc = could not find container \"832b6d3b6d35484729f6a4775af16b4793c353fd00defe20bf1a5464b1df2945\": container with ID starting with 832b6d3b6d35484729f6a4775af16b4793c353fd00defe20bf1a5464b1df2945 not found: ID does not exist" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.522122 4578 scope.go:117] "RemoveContainer" containerID="e58105bd2a630f9aa5ec1ad69fcd66bec0bec5aa10ad13a81309014b56d280c9" Oct 03 13:15:30 crc kubenswrapper[4578]: E1003 13:15:30.522410 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e58105bd2a630f9aa5ec1ad69fcd66bec0bec5aa10ad13a81309014b56d280c9\": container with ID starting with e58105bd2a630f9aa5ec1ad69fcd66bec0bec5aa10ad13a81309014b56d280c9 not found: ID does not exist" containerID="e58105bd2a630f9aa5ec1ad69fcd66bec0bec5aa10ad13a81309014b56d280c9" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.522431 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e58105bd2a630f9aa5ec1ad69fcd66bec0bec5aa10ad13a81309014b56d280c9"} err="failed to get container status \"e58105bd2a630f9aa5ec1ad69fcd66bec0bec5aa10ad13a81309014b56d280c9\": rpc error: code = NotFound desc = could not find container \"e58105bd2a630f9aa5ec1ad69fcd66bec0bec5aa10ad13a81309014b56d280c9\": container with ID starting with e58105bd2a630f9aa5ec1ad69fcd66bec0bec5aa10ad13a81309014b56d280c9 not found: ID does not exist" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.522447 4578 scope.go:117] "RemoveContainer" containerID="c9e3ccec20bb8a4d4eb4b682d12a2bd984b35012a2dc2c2f97cdbed39edee63a" Oct 03 13:15:30 crc kubenswrapper[4578]: E1003 13:15:30.522840 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9e3ccec20bb8a4d4eb4b682d12a2bd984b35012a2dc2c2f97cdbed39edee63a\": container with ID starting with c9e3ccec20bb8a4d4eb4b682d12a2bd984b35012a2dc2c2f97cdbed39edee63a not found: ID does not exist" containerID="c9e3ccec20bb8a4d4eb4b682d12a2bd984b35012a2dc2c2f97cdbed39edee63a" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.522862 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9e3ccec20bb8a4d4eb4b682d12a2bd984b35012a2dc2c2f97cdbed39edee63a"} err="failed to get container status \"c9e3ccec20bb8a4d4eb4b682d12a2bd984b35012a2dc2c2f97cdbed39edee63a\": rpc error: code = NotFound desc = could not find container \"c9e3ccec20bb8a4d4eb4b682d12a2bd984b35012a2dc2c2f97cdbed39edee63a\": container with ID starting with c9e3ccec20bb8a4d4eb4b682d12a2bd984b35012a2dc2c2f97cdbed39edee63a not found: ID does not exist" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.658213 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.658423 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0edcba27-85bd-45cd-9eb9-2759fa4b07ca" containerName="nova-api-log" containerID="cri-o://8c5c130602f259d06dc208f8e931341ec87ddd4840dad5aa40a13a0f03673f83" gracePeriod=30 Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.658493 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0edcba27-85bd-45cd-9eb9-2759fa4b07ca" containerName="nova-api-api" containerID="cri-o://242d22f7ce31b5ac25ceff2ed17b5b4522e4a5b51777882a8a630cfceda49eaf" gracePeriod=30 Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.741920 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.742023 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 13:15:30 crc kubenswrapper[4578]: I1003 13:15:30.920240 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5" path="/var/lib/kubelet/pods/0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5/volumes" Oct 03 13:15:31 crc kubenswrapper[4578]: I1003 13:15:31.000912 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:15:31 crc kubenswrapper[4578]: I1003 13:15:31.001251 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerName="ceilometer-central-agent" containerID="cri-o://7a24c55735edc8b55d25b89a091ef7ca3d644e8a1affa5bd1b699876fb8b363e" gracePeriod=30 Oct 03 13:15:31 crc kubenswrapper[4578]: I1003 13:15:31.001339 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerName="proxy-httpd" containerID="cri-o://a7147cc315d8b21fbd71f93362f1f57c969862885b1f0de7294648a8c5016cb8" gracePeriod=30 Oct 03 13:15:31 crc kubenswrapper[4578]: I1003 13:15:31.001378 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerName="sg-core" containerID="cri-o://1e2850122657463934810cf8beaa25b0b49f53dd5ba5a3c0c3b1b31b9797a6a1" gracePeriod=30 Oct 03 13:15:31 crc kubenswrapper[4578]: I1003 13:15:31.001432 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerName="ceilometer-notification-agent" containerID="cri-o://239d72981276c68c743a6f7b7a0e09e38b2b12cec6fa41760cf8b5e459e6f02a" gracePeriod=30 Oct 03 13:15:31 crc kubenswrapper[4578]: I1003 13:15:31.040598 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.194:3000/\": read tcp 10.217.0.2:39670->10.217.0.194:3000: read: connection reset by peer" Oct 03 13:15:31 crc kubenswrapper[4578]: I1003 13:15:31.433355 4578 generic.go:334] "Generic (PLEG): container finished" podID="0edcba27-85bd-45cd-9eb9-2759fa4b07ca" containerID="8c5c130602f259d06dc208f8e931341ec87ddd4840dad5aa40a13a0f03673f83" exitCode=143 Oct 03 13:15:31 crc kubenswrapper[4578]: I1003 13:15:31.433479 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0edcba27-85bd-45cd-9eb9-2759fa4b07ca","Type":"ContainerDied","Data":"8c5c130602f259d06dc208f8e931341ec87ddd4840dad5aa40a13a0f03673f83"} Oct 03 13:15:31 crc kubenswrapper[4578]: I1003 13:15:31.436662 4578 generic.go:334] "Generic (PLEG): container finished" podID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerID="a7147cc315d8b21fbd71f93362f1f57c969862885b1f0de7294648a8c5016cb8" exitCode=0 Oct 03 13:15:31 crc kubenswrapper[4578]: I1003 13:15:31.436690 4578 generic.go:334] "Generic (PLEG): container finished" podID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerID="1e2850122657463934810cf8beaa25b0b49f53dd5ba5a3c0c3b1b31b9797a6a1" exitCode=2 Oct 03 13:15:31 crc kubenswrapper[4578]: I1003 13:15:31.436743 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d8e8cfc9-5761-4caf-9702-a0809a596754","Type":"ContainerDied","Data":"a7147cc315d8b21fbd71f93362f1f57c969862885b1f0de7294648a8c5016cb8"} Oct 03 13:15:31 crc kubenswrapper[4578]: I1003 13:15:31.436796 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d8e8cfc9-5761-4caf-9702-a0809a596754","Type":"ContainerDied","Data":"1e2850122657463934810cf8beaa25b0b49f53dd5ba5a3c0c3b1b31b9797a6a1"} Oct 03 13:15:32 crc kubenswrapper[4578]: I1003 13:15:32.450695 4578 generic.go:334] "Generic (PLEG): container finished" podID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerID="7a24c55735edc8b55d25b89a091ef7ca3d644e8a1affa5bd1b699876fb8b363e" exitCode=0 Oct 03 13:15:32 crc kubenswrapper[4578]: I1003 13:15:32.450764 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d8e8cfc9-5761-4caf-9702-a0809a596754","Type":"ContainerDied","Data":"7a24c55735edc8b55d25b89a091ef7ca3d644e8a1affa5bd1b699876fb8b363e"} Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.317036 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.452578 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-config-data\") pod \"0edcba27-85bd-45cd-9eb9-2759fa4b07ca\" (UID: \"0edcba27-85bd-45cd-9eb9-2759fa4b07ca\") " Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.452883 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-logs\") pod \"0edcba27-85bd-45cd-9eb9-2759fa4b07ca\" (UID: \"0edcba27-85bd-45cd-9eb9-2759fa4b07ca\") " Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.453002 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfwpm\" (UniqueName: \"kubernetes.io/projected/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-kube-api-access-xfwpm\") pod \"0edcba27-85bd-45cd-9eb9-2759fa4b07ca\" (UID: \"0edcba27-85bd-45cd-9eb9-2759fa4b07ca\") " Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.453079 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-combined-ca-bundle\") pod \"0edcba27-85bd-45cd-9eb9-2759fa4b07ca\" (UID: \"0edcba27-85bd-45cd-9eb9-2759fa4b07ca\") " Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.454000 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-logs" (OuterVolumeSpecName: "logs") pod "0edcba27-85bd-45cd-9eb9-2759fa4b07ca" (UID: "0edcba27-85bd-45cd-9eb9-2759fa4b07ca"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.454222 4578 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.491488 4578 generic.go:334] "Generic (PLEG): container finished" podID="0edcba27-85bd-45cd-9eb9-2759fa4b07ca" containerID="242d22f7ce31b5ac25ceff2ed17b5b4522e4a5b51777882a8a630cfceda49eaf" exitCode=0 Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.491544 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.491562 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0edcba27-85bd-45cd-9eb9-2759fa4b07ca","Type":"ContainerDied","Data":"242d22f7ce31b5ac25ceff2ed17b5b4522e4a5b51777882a8a630cfceda49eaf"} Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.491891 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0edcba27-85bd-45cd-9eb9-2759fa4b07ca","Type":"ContainerDied","Data":"3e3a565610e85710f456133e46b8e4357b45a85c276073665e1a02e035be198a"} Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.491908 4578 scope.go:117] "RemoveContainer" containerID="242d22f7ce31b5ac25ceff2ed17b5b4522e4a5b51777882a8a630cfceda49eaf" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.517189 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-kube-api-access-xfwpm" (OuterVolumeSpecName: "kube-api-access-xfwpm") pod "0edcba27-85bd-45cd-9eb9-2759fa4b07ca" (UID: "0edcba27-85bd-45cd-9eb9-2759fa4b07ca"). InnerVolumeSpecName "kube-api-access-xfwpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.534838 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-config-data" (OuterVolumeSpecName: "config-data") pod "0edcba27-85bd-45cd-9eb9-2759fa4b07ca" (UID: "0edcba27-85bd-45cd-9eb9-2759fa4b07ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.553799 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0edcba27-85bd-45cd-9eb9-2759fa4b07ca" (UID: "0edcba27-85bd-45cd-9eb9-2759fa4b07ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.557992 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfwpm\" (UniqueName: \"kubernetes.io/projected/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-kube-api-access-xfwpm\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.558017 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.558027 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0edcba27-85bd-45cd-9eb9-2759fa4b07ca-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.627921 4578 scope.go:117] "RemoveContainer" containerID="8c5c130602f259d06dc208f8e931341ec87ddd4840dad5aa40a13a0f03673f83" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.659472 4578 scope.go:117] "RemoveContainer" containerID="242d22f7ce31b5ac25ceff2ed17b5b4522e4a5b51777882a8a630cfceda49eaf" Oct 03 13:15:34 crc kubenswrapper[4578]: E1003 13:15:34.660404 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"242d22f7ce31b5ac25ceff2ed17b5b4522e4a5b51777882a8a630cfceda49eaf\": container with ID starting with 242d22f7ce31b5ac25ceff2ed17b5b4522e4a5b51777882a8a630cfceda49eaf not found: ID does not exist" containerID="242d22f7ce31b5ac25ceff2ed17b5b4522e4a5b51777882a8a630cfceda49eaf" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.660468 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"242d22f7ce31b5ac25ceff2ed17b5b4522e4a5b51777882a8a630cfceda49eaf"} err="failed to get container status \"242d22f7ce31b5ac25ceff2ed17b5b4522e4a5b51777882a8a630cfceda49eaf\": rpc error: code = NotFound desc = could not find container \"242d22f7ce31b5ac25ceff2ed17b5b4522e4a5b51777882a8a630cfceda49eaf\": container with ID starting with 242d22f7ce31b5ac25ceff2ed17b5b4522e4a5b51777882a8a630cfceda49eaf not found: ID does not exist" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.660515 4578 scope.go:117] "RemoveContainer" containerID="8c5c130602f259d06dc208f8e931341ec87ddd4840dad5aa40a13a0f03673f83" Oct 03 13:15:34 crc kubenswrapper[4578]: E1003 13:15:34.661134 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c5c130602f259d06dc208f8e931341ec87ddd4840dad5aa40a13a0f03673f83\": container with ID starting with 8c5c130602f259d06dc208f8e931341ec87ddd4840dad5aa40a13a0f03673f83 not found: ID does not exist" containerID="8c5c130602f259d06dc208f8e931341ec87ddd4840dad5aa40a13a0f03673f83" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.661188 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c5c130602f259d06dc208f8e931341ec87ddd4840dad5aa40a13a0f03673f83"} err="failed to get container status \"8c5c130602f259d06dc208f8e931341ec87ddd4840dad5aa40a13a0f03673f83\": rpc error: code = NotFound desc = could not find container \"8c5c130602f259d06dc208f8e931341ec87ddd4840dad5aa40a13a0f03673f83\": container with ID starting with 8c5c130602f259d06dc208f8e931341ec87ddd4840dad5aa40a13a0f03673f83 not found: ID does not exist" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.868737 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.878462 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.891725 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 03 13:15:34 crc kubenswrapper[4578]: E1003 13:15:34.893312 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5" containerName="extract-content" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.894111 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5" containerName="extract-content" Oct 03 13:15:34 crc kubenswrapper[4578]: E1003 13:15:34.894162 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0edcba27-85bd-45cd-9eb9-2759fa4b07ca" containerName="nova-api-api" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.894169 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="0edcba27-85bd-45cd-9eb9-2759fa4b07ca" containerName="nova-api-api" Oct 03 13:15:34 crc kubenswrapper[4578]: E1003 13:15:34.894182 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5" containerName="extract-utilities" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.894187 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5" containerName="extract-utilities" Oct 03 13:15:34 crc kubenswrapper[4578]: E1003 13:15:34.894196 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5" containerName="registry-server" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.894202 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5" containerName="registry-server" Oct 03 13:15:34 crc kubenswrapper[4578]: E1003 13:15:34.894223 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0edcba27-85bd-45cd-9eb9-2759fa4b07ca" containerName="nova-api-log" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.894229 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="0edcba27-85bd-45cd-9eb9-2759fa4b07ca" containerName="nova-api-log" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.894558 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="0edcba27-85bd-45cd-9eb9-2759fa4b07ca" containerName="nova-api-api" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.894576 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="0edcba27-85bd-45cd-9eb9-2759fa4b07ca" containerName="nova-api-log" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.894588 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="0602b27a-ab33-4bbb-a5cd-9473bfb3ddd5" containerName="registry-server" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.895614 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.900018 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.903687 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.904470 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.965560 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0edcba27-85bd-45cd-9eb9-2759fa4b07ca" path="/var/lib/kubelet/pods/0edcba27-85bd-45cd-9eb9-2759fa4b07ca/volumes" Oct 03 13:15:34 crc kubenswrapper[4578]: I1003 13:15:34.966193 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.014953 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.047578 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.077888 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " pod="openstack/nova-api-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.077997 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-config-data\") pod \"nova-api-0\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " pod="openstack/nova-api-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.078133 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-logs\") pod \"nova-api-0\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " pod="openstack/nova-api-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.078151 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vqdp\" (UniqueName: \"kubernetes.io/projected/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-kube-api-access-6vqdp\") pod \"nova-api-0\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " pod="openstack/nova-api-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.078183 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-public-tls-certs\") pod \"nova-api-0\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " pod="openstack/nova-api-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.078214 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " pod="openstack/nova-api-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.091967 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.092030 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.145775 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.179362 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " pod="openstack/nova-api-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.179420 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-config-data\") pod \"nova-api-0\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " pod="openstack/nova-api-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.179488 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-logs\") pod \"nova-api-0\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " pod="openstack/nova-api-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.179503 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vqdp\" (UniqueName: \"kubernetes.io/projected/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-kube-api-access-6vqdp\") pod \"nova-api-0\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " pod="openstack/nova-api-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.179526 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-public-tls-certs\") pod \"nova-api-0\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " pod="openstack/nova-api-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.179559 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " pod="openstack/nova-api-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.185105 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-logs\") pod \"nova-api-0\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " pod="openstack/nova-api-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.187356 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-public-tls-certs\") pod \"nova-api-0\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " pod="openstack/nova-api-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.188005 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " pod="openstack/nova-api-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.189895 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " pod="openstack/nova-api-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.203441 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vqdp\" (UniqueName: \"kubernetes.io/projected/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-kube-api-access-6vqdp\") pod \"nova-api-0\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " pod="openstack/nova-api-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.207517 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-config-data\") pod \"nova-api-0\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " pod="openstack/nova-api-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.275308 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.282243 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-combined-ca-bundle\") pod \"d8e8cfc9-5761-4caf-9702-a0809a596754\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.282292 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-scripts\") pod \"d8e8cfc9-5761-4caf-9702-a0809a596754\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.282316 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qzps\" (UniqueName: \"kubernetes.io/projected/d8e8cfc9-5761-4caf-9702-a0809a596754-kube-api-access-9qzps\") pod \"d8e8cfc9-5761-4caf-9702-a0809a596754\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.282362 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d8e8cfc9-5761-4caf-9702-a0809a596754-log-httpd\") pod \"d8e8cfc9-5761-4caf-9702-a0809a596754\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.282426 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-sg-core-conf-yaml\") pod \"d8e8cfc9-5761-4caf-9702-a0809a596754\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.282471 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-ceilometer-tls-certs\") pod \"d8e8cfc9-5761-4caf-9702-a0809a596754\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.282596 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d8e8cfc9-5761-4caf-9702-a0809a596754-run-httpd\") pod \"d8e8cfc9-5761-4caf-9702-a0809a596754\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.282870 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-config-data\") pod \"d8e8cfc9-5761-4caf-9702-a0809a596754\" (UID: \"d8e8cfc9-5761-4caf-9702-a0809a596754\") " Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.283088 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8e8cfc9-5761-4caf-9702-a0809a596754-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d8e8cfc9-5761-4caf-9702-a0809a596754" (UID: "d8e8cfc9-5761-4caf-9702-a0809a596754"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.283108 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8e8cfc9-5761-4caf-9702-a0809a596754-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d8e8cfc9-5761-4caf-9702-a0809a596754" (UID: "d8e8cfc9-5761-4caf-9702-a0809a596754"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.284190 4578 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d8e8cfc9-5761-4caf-9702-a0809a596754-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.293326 4578 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d8e8cfc9-5761-4caf-9702-a0809a596754-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.287767 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8e8cfc9-5761-4caf-9702-a0809a596754-kube-api-access-9qzps" (OuterVolumeSpecName: "kube-api-access-9qzps") pod "d8e8cfc9-5761-4caf-9702-a0809a596754" (UID: "d8e8cfc9-5761-4caf-9702-a0809a596754"). InnerVolumeSpecName "kube-api-access-9qzps". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.296011 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-scripts" (OuterVolumeSpecName: "scripts") pod "d8e8cfc9-5761-4caf-9702-a0809a596754" (UID: "d8e8cfc9-5761-4caf-9702-a0809a596754"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.325391 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d8e8cfc9-5761-4caf-9702-a0809a596754" (UID: "d8e8cfc9-5761-4caf-9702-a0809a596754"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.351954 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "d8e8cfc9-5761-4caf-9702-a0809a596754" (UID: "d8e8cfc9-5761-4caf-9702-a0809a596754"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.396966 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.396996 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qzps\" (UniqueName: \"kubernetes.io/projected/d8e8cfc9-5761-4caf-9702-a0809a596754-kube-api-access-9qzps\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.397005 4578 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.397014 4578 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.406967 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d8e8cfc9-5761-4caf-9702-a0809a596754" (UID: "d8e8cfc9-5761-4caf-9702-a0809a596754"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.435958 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-config-data" (OuterVolumeSpecName: "config-data") pod "d8e8cfc9-5761-4caf-9702-a0809a596754" (UID: "d8e8cfc9-5761-4caf-9702-a0809a596754"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.498663 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.498878 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8e8cfc9-5761-4caf-9702-a0809a596754-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.501522 4578 generic.go:334] "Generic (PLEG): container finished" podID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerID="239d72981276c68c743a6f7b7a0e09e38b2b12cec6fa41760cf8b5e459e6f02a" exitCode=0 Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.501705 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d8e8cfc9-5761-4caf-9702-a0809a596754","Type":"ContainerDied","Data":"239d72981276c68c743a6f7b7a0e09e38b2b12cec6fa41760cf8b5e459e6f02a"} Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.501851 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d8e8cfc9-5761-4caf-9702-a0809a596754","Type":"ContainerDied","Data":"d5576c016acdfd1becb1f21311dbb08674d2ad63ab4ab4bb87170462d679ac60"} Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.501919 4578 scope.go:117] "RemoveContainer" containerID="a7147cc315d8b21fbd71f93362f1f57c969862885b1f0de7294648a8c5016cb8" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.501924 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.533760 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.539120 4578 scope.go:117] "RemoveContainer" containerID="1e2850122657463934810cf8beaa25b0b49f53dd5ba5a3c0c3b1b31b9797a6a1" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.552929 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.563969 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.578953 4578 scope.go:117] "RemoveContainer" containerID="239d72981276c68c743a6f7b7a0e09e38b2b12cec6fa41760cf8b5e459e6f02a" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.582797 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:15:35 crc kubenswrapper[4578]: E1003 13:15:35.583362 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerName="ceilometer-central-agent" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.583385 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerName="ceilometer-central-agent" Oct 03 13:15:35 crc kubenswrapper[4578]: E1003 13:15:35.583398 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerName="proxy-httpd" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.583408 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerName="proxy-httpd" Oct 03 13:15:35 crc kubenswrapper[4578]: E1003 13:15:35.583421 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerName="ceilometer-notification-agent" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.583429 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerName="ceilometer-notification-agent" Oct 03 13:15:35 crc kubenswrapper[4578]: E1003 13:15:35.583462 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerName="sg-core" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.583470 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerName="sg-core" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.583708 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerName="proxy-httpd" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.583738 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerName="sg-core" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.583751 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerName="ceilometer-notification-agent" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.583775 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8e8cfc9-5761-4caf-9702-a0809a596754" containerName="ceilometer-central-agent" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.586043 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.589930 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.590699 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.591050 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.615911 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.625772 4578 scope.go:117] "RemoveContainer" containerID="7a24c55735edc8b55d25b89a091ef7ca3d644e8a1affa5bd1b699876fb8b363e" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.659370 4578 scope.go:117] "RemoveContainer" containerID="a7147cc315d8b21fbd71f93362f1f57c969862885b1f0de7294648a8c5016cb8" Oct 03 13:15:35 crc kubenswrapper[4578]: E1003 13:15:35.659958 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7147cc315d8b21fbd71f93362f1f57c969862885b1f0de7294648a8c5016cb8\": container with ID starting with a7147cc315d8b21fbd71f93362f1f57c969862885b1f0de7294648a8c5016cb8 not found: ID does not exist" containerID="a7147cc315d8b21fbd71f93362f1f57c969862885b1f0de7294648a8c5016cb8" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.660002 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7147cc315d8b21fbd71f93362f1f57c969862885b1f0de7294648a8c5016cb8"} err="failed to get container status \"a7147cc315d8b21fbd71f93362f1f57c969862885b1f0de7294648a8c5016cb8\": rpc error: code = NotFound desc = could not find container \"a7147cc315d8b21fbd71f93362f1f57c969862885b1f0de7294648a8c5016cb8\": container with ID starting with a7147cc315d8b21fbd71f93362f1f57c969862885b1f0de7294648a8c5016cb8 not found: ID does not exist" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.660034 4578 scope.go:117] "RemoveContainer" containerID="1e2850122657463934810cf8beaa25b0b49f53dd5ba5a3c0c3b1b31b9797a6a1" Oct 03 13:15:35 crc kubenswrapper[4578]: E1003 13:15:35.660909 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e2850122657463934810cf8beaa25b0b49f53dd5ba5a3c0c3b1b31b9797a6a1\": container with ID starting with 1e2850122657463934810cf8beaa25b0b49f53dd5ba5a3c0c3b1b31b9797a6a1 not found: ID does not exist" containerID="1e2850122657463934810cf8beaa25b0b49f53dd5ba5a3c0c3b1b31b9797a6a1" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.660971 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e2850122657463934810cf8beaa25b0b49f53dd5ba5a3c0c3b1b31b9797a6a1"} err="failed to get container status \"1e2850122657463934810cf8beaa25b0b49f53dd5ba5a3c0c3b1b31b9797a6a1\": rpc error: code = NotFound desc = could not find container \"1e2850122657463934810cf8beaa25b0b49f53dd5ba5a3c0c3b1b31b9797a6a1\": container with ID starting with 1e2850122657463934810cf8beaa25b0b49f53dd5ba5a3c0c3b1b31b9797a6a1 not found: ID does not exist" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.660997 4578 scope.go:117] "RemoveContainer" containerID="239d72981276c68c743a6f7b7a0e09e38b2b12cec6fa41760cf8b5e459e6f02a" Oct 03 13:15:35 crc kubenswrapper[4578]: E1003 13:15:35.661736 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"239d72981276c68c743a6f7b7a0e09e38b2b12cec6fa41760cf8b5e459e6f02a\": container with ID starting with 239d72981276c68c743a6f7b7a0e09e38b2b12cec6fa41760cf8b5e459e6f02a not found: ID does not exist" containerID="239d72981276c68c743a6f7b7a0e09e38b2b12cec6fa41760cf8b5e459e6f02a" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.661770 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"239d72981276c68c743a6f7b7a0e09e38b2b12cec6fa41760cf8b5e459e6f02a"} err="failed to get container status \"239d72981276c68c743a6f7b7a0e09e38b2b12cec6fa41760cf8b5e459e6f02a\": rpc error: code = NotFound desc = could not find container \"239d72981276c68c743a6f7b7a0e09e38b2b12cec6fa41760cf8b5e459e6f02a\": container with ID starting with 239d72981276c68c743a6f7b7a0e09e38b2b12cec6fa41760cf8b5e459e6f02a not found: ID does not exist" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.661795 4578 scope.go:117] "RemoveContainer" containerID="7a24c55735edc8b55d25b89a091ef7ca3d644e8a1affa5bd1b699876fb8b363e" Oct 03 13:15:35 crc kubenswrapper[4578]: E1003 13:15:35.668805 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a24c55735edc8b55d25b89a091ef7ca3d644e8a1affa5bd1b699876fb8b363e\": container with ID starting with 7a24c55735edc8b55d25b89a091ef7ca3d644e8a1affa5bd1b699876fb8b363e not found: ID does not exist" containerID="7a24c55735edc8b55d25b89a091ef7ca3d644e8a1affa5bd1b699876fb8b363e" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.668860 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a24c55735edc8b55d25b89a091ef7ca3d644e8a1affa5bd1b699876fb8b363e"} err="failed to get container status \"7a24c55735edc8b55d25b89a091ef7ca3d644e8a1affa5bd1b699876fb8b363e\": rpc error: code = NotFound desc = could not find container \"7a24c55735edc8b55d25b89a091ef7ca3d644e8a1affa5bd1b699876fb8b363e\": container with ID starting with 7a24c55735edc8b55d25b89a091ef7ca3d644e8a1affa5bd1b699876fb8b363e not found: ID does not exist" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.702891 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkfb5\" (UniqueName: \"kubernetes.io/projected/a54925d7-6f48-4ce4-8062-c81a33abb815-kube-api-access-rkfb5\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.702957 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a54925d7-6f48-4ce4-8062-c81a33abb815-run-httpd\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.702989 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a54925d7-6f48-4ce4-8062-c81a33abb815-config-data\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.703009 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a54925d7-6f48-4ce4-8062-c81a33abb815-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.703025 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a54925d7-6f48-4ce4-8062-c81a33abb815-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.703231 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a54925d7-6f48-4ce4-8062-c81a33abb815-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.703257 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a54925d7-6f48-4ce4-8062-c81a33abb815-scripts\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.703281 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a54925d7-6f48-4ce4-8062-c81a33abb815-log-httpd\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.734487 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-mz6wt"] Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.739153 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mz6wt" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.741551 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.741574 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.742551 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.744881 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.761741 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-mz6wt"] Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.768925 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.806032 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a54925d7-6f48-4ce4-8062-c81a33abb815-run-httpd\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.806418 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a54925d7-6f48-4ce4-8062-c81a33abb815-run-httpd\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.806508 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a54925d7-6f48-4ce4-8062-c81a33abb815-config-data\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.807236 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a54925d7-6f48-4ce4-8062-c81a33abb815-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.807266 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a54925d7-6f48-4ce4-8062-c81a33abb815-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.807910 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a54925d7-6f48-4ce4-8062-c81a33abb815-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.807937 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a54925d7-6f48-4ce4-8062-c81a33abb815-scripts\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.807951 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a54925d7-6f48-4ce4-8062-c81a33abb815-log-httpd\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.808023 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkfb5\" (UniqueName: \"kubernetes.io/projected/a54925d7-6f48-4ce4-8062-c81a33abb815-kube-api-access-rkfb5\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.812243 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a54925d7-6f48-4ce4-8062-c81a33abb815-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.812447 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a54925d7-6f48-4ce4-8062-c81a33abb815-log-httpd\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.813055 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a54925d7-6f48-4ce4-8062-c81a33abb815-config-data\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.813653 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a54925d7-6f48-4ce4-8062-c81a33abb815-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.816886 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a54925d7-6f48-4ce4-8062-c81a33abb815-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.818878 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a54925d7-6f48-4ce4-8062-c81a33abb815-scripts\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.827244 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkfb5\" (UniqueName: \"kubernetes.io/projected/a54925d7-6f48-4ce4-8062-c81a33abb815-kube-api-access-rkfb5\") pod \"ceilometer-0\" (UID: \"a54925d7-6f48-4ce4-8062-c81a33abb815\") " pod="openstack/ceilometer-0" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.909572 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3e10cb8-af89-4561-a4ae-804943cf9440-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mz6wt\" (UID: \"d3e10cb8-af89-4561-a4ae-804943cf9440\") " pod="openstack/nova-cell1-cell-mapping-mz6wt" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.909677 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3e10cb8-af89-4561-a4ae-804943cf9440-scripts\") pod \"nova-cell1-cell-mapping-mz6wt\" (UID: \"d3e10cb8-af89-4561-a4ae-804943cf9440\") " pod="openstack/nova-cell1-cell-mapping-mz6wt" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.909715 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3e10cb8-af89-4561-a4ae-804943cf9440-config-data\") pod \"nova-cell1-cell-mapping-mz6wt\" (UID: \"d3e10cb8-af89-4561-a4ae-804943cf9440\") " pod="openstack/nova-cell1-cell-mapping-mz6wt" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.909773 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llp7r\" (UniqueName: \"kubernetes.io/projected/d3e10cb8-af89-4561-a4ae-804943cf9440-kube-api-access-llp7r\") pod \"nova-cell1-cell-mapping-mz6wt\" (UID: \"d3e10cb8-af89-4561-a4ae-804943cf9440\") " pod="openstack/nova-cell1-cell-mapping-mz6wt" Oct 03 13:15:35 crc kubenswrapper[4578]: I1003 13:15:35.921864 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 03 13:15:36 crc kubenswrapper[4578]: I1003 13:15:36.011840 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llp7r\" (UniqueName: \"kubernetes.io/projected/d3e10cb8-af89-4561-a4ae-804943cf9440-kube-api-access-llp7r\") pod \"nova-cell1-cell-mapping-mz6wt\" (UID: \"d3e10cb8-af89-4561-a4ae-804943cf9440\") " pod="openstack/nova-cell1-cell-mapping-mz6wt" Oct 03 13:15:36 crc kubenswrapper[4578]: I1003 13:15:36.011942 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3e10cb8-af89-4561-a4ae-804943cf9440-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mz6wt\" (UID: \"d3e10cb8-af89-4561-a4ae-804943cf9440\") " pod="openstack/nova-cell1-cell-mapping-mz6wt" Oct 03 13:15:36 crc kubenswrapper[4578]: I1003 13:15:36.012018 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3e10cb8-af89-4561-a4ae-804943cf9440-scripts\") pod \"nova-cell1-cell-mapping-mz6wt\" (UID: \"d3e10cb8-af89-4561-a4ae-804943cf9440\") " pod="openstack/nova-cell1-cell-mapping-mz6wt" Oct 03 13:15:36 crc kubenswrapper[4578]: I1003 13:15:36.012058 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3e10cb8-af89-4561-a4ae-804943cf9440-config-data\") pod \"nova-cell1-cell-mapping-mz6wt\" (UID: \"d3e10cb8-af89-4561-a4ae-804943cf9440\") " pod="openstack/nova-cell1-cell-mapping-mz6wt" Oct 03 13:15:36 crc kubenswrapper[4578]: I1003 13:15:36.020416 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3e10cb8-af89-4561-a4ae-804943cf9440-scripts\") pod \"nova-cell1-cell-mapping-mz6wt\" (UID: \"d3e10cb8-af89-4561-a4ae-804943cf9440\") " pod="openstack/nova-cell1-cell-mapping-mz6wt" Oct 03 13:15:36 crc kubenswrapper[4578]: I1003 13:15:36.020662 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3e10cb8-af89-4561-a4ae-804943cf9440-config-data\") pod \"nova-cell1-cell-mapping-mz6wt\" (UID: \"d3e10cb8-af89-4561-a4ae-804943cf9440\") " pod="openstack/nova-cell1-cell-mapping-mz6wt" Oct 03 13:15:36 crc kubenswrapper[4578]: I1003 13:15:36.021196 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3e10cb8-af89-4561-a4ae-804943cf9440-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mz6wt\" (UID: \"d3e10cb8-af89-4561-a4ae-804943cf9440\") " pod="openstack/nova-cell1-cell-mapping-mz6wt" Oct 03 13:15:36 crc kubenswrapper[4578]: I1003 13:15:36.033048 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llp7r\" (UniqueName: \"kubernetes.io/projected/d3e10cb8-af89-4561-a4ae-804943cf9440-kube-api-access-llp7r\") pod \"nova-cell1-cell-mapping-mz6wt\" (UID: \"d3e10cb8-af89-4561-a4ae-804943cf9440\") " pod="openstack/nova-cell1-cell-mapping-mz6wt" Oct 03 13:15:36 crc kubenswrapper[4578]: I1003 13:15:36.078740 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mz6wt" Oct 03 13:15:36 crc kubenswrapper[4578]: I1003 13:15:36.509499 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 03 13:15:36 crc kubenswrapper[4578]: I1003 13:15:36.547491 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aa56cf15-10a3-4978-a7cd-1d0c32d1377c","Type":"ContainerStarted","Data":"f9ee50c58c5c748dd8c43e06176ea153e9a4e6056f78a3923f9a4c2d9b9fe25d"} Oct 03 13:15:36 crc kubenswrapper[4578]: I1003 13:15:36.547560 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aa56cf15-10a3-4978-a7cd-1d0c32d1377c","Type":"ContainerStarted","Data":"cf0c720e06b7cd264814a0a477a88a4be612b3c3a32536da0b8dd3b3c7ea5a8d"} Oct 03 13:15:36 crc kubenswrapper[4578]: I1003 13:15:36.786621 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-mz6wt"] Oct 03 13:15:36 crc kubenswrapper[4578]: I1003 13:15:36.832990 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="34b83c0e-8c44-4ceb-abae-35551eb3f183" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:15:36 crc kubenswrapper[4578]: I1003 13:15:36.833016 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="34b83c0e-8c44-4ceb-abae-35551eb3f183" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:15:36 crc kubenswrapper[4578]: I1003 13:15:36.963321 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8e8cfc9-5761-4caf-9702-a0809a596754" path="/var/lib/kubelet/pods/d8e8cfc9-5761-4caf-9702-a0809a596754/volumes" Oct 03 13:15:37 crc kubenswrapper[4578]: I1003 13:15:37.559574 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aa56cf15-10a3-4978-a7cd-1d0c32d1377c","Type":"ContainerStarted","Data":"afaed3a1fb21a22bb87ac250d1cbd6ff3af9bbe14f49dc554001442d1544a936"} Oct 03 13:15:37 crc kubenswrapper[4578]: I1003 13:15:37.561890 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a54925d7-6f48-4ce4-8062-c81a33abb815","Type":"ContainerStarted","Data":"30b7dc611c1bb15570f8524b0e503b836b3415966b365159d6b62295968e44d4"} Oct 03 13:15:37 crc kubenswrapper[4578]: I1003 13:15:37.561937 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a54925d7-6f48-4ce4-8062-c81a33abb815","Type":"ContainerStarted","Data":"c7f18fc29206a89a22df97843543fb5ec42e19d606d25a2bd356774a550e0b01"} Oct 03 13:15:37 crc kubenswrapper[4578]: I1003 13:15:37.563680 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mz6wt" event={"ID":"d3e10cb8-af89-4561-a4ae-804943cf9440","Type":"ContainerStarted","Data":"c4fd7d19b6ab5bea908d1b9c1dcde5509c71ba22e94c933a286e2563297c1ef8"} Oct 03 13:15:37 crc kubenswrapper[4578]: I1003 13:15:37.563779 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mz6wt" event={"ID":"d3e10cb8-af89-4561-a4ae-804943cf9440","Type":"ContainerStarted","Data":"e0a4e8584b26ffe81f3dc9bc5a49109bbf05a323a946355cc3e43b56e2c9ad4d"} Oct 03 13:15:37 crc kubenswrapper[4578]: I1003 13:15:37.589668 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.5896484810000002 podStartE2EDuration="3.589648481s" podCreationTimestamp="2025-10-03 13:15:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:15:37.584032555 +0000 UTC m=+1473.382504739" watchObservedRunningTime="2025-10-03 13:15:37.589648481 +0000 UTC m=+1473.388120665" Oct 03 13:15:37 crc kubenswrapper[4578]: I1003 13:15:37.600512 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-mz6wt" podStartSLOduration=2.600498431 podStartE2EDuration="2.600498431s" podCreationTimestamp="2025-10-03 13:15:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:15:37.596544548 +0000 UTC m=+1473.395016732" watchObservedRunningTime="2025-10-03 13:15:37.600498431 +0000 UTC m=+1473.398970615" Oct 03 13:15:37 crc kubenswrapper[4578]: I1003 13:15:37.992806 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:15:38 crc kubenswrapper[4578]: I1003 13:15:38.061320 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-g7kjh"] Oct 03 13:15:38 crc kubenswrapper[4578]: I1003 13:15:38.061810 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" podUID="81676263-327d-4d19-8673-a44d5e576430" containerName="dnsmasq-dns" containerID="cri-o://4b1b34611cc3274a88742b81445a6c0ca5c68f7a7c316eceb9dad6764ab43967" gracePeriod=10 Oct 03 13:15:38 crc kubenswrapper[4578]: I1003 13:15:38.631992 4578 generic.go:334] "Generic (PLEG): container finished" podID="81676263-327d-4d19-8673-a44d5e576430" containerID="4b1b34611cc3274a88742b81445a6c0ca5c68f7a7c316eceb9dad6764ab43967" exitCode=0 Oct 03 13:15:38 crc kubenswrapper[4578]: I1003 13:15:38.632601 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" event={"ID":"81676263-327d-4d19-8673-a44d5e576430","Type":"ContainerDied","Data":"4b1b34611cc3274a88742b81445a6c0ca5c68f7a7c316eceb9dad6764ab43967"} Oct 03 13:15:38 crc kubenswrapper[4578]: I1003 13:15:38.653470 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a54925d7-6f48-4ce4-8062-c81a33abb815","Type":"ContainerStarted","Data":"fbc7a6119571015a742d0a46af1d2575c9fb9c285f22c1a26a4818cd4146009e"} Oct 03 13:15:38 crc kubenswrapper[4578]: I1003 13:15:38.728920 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:15:38 crc kubenswrapper[4578]: I1003 13:15:38.818182 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-ovsdbserver-nb\") pod \"81676263-327d-4d19-8673-a44d5e576430\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " Oct 03 13:15:38 crc kubenswrapper[4578]: I1003 13:15:38.818459 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-config\") pod \"81676263-327d-4d19-8673-a44d5e576430\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " Oct 03 13:15:38 crc kubenswrapper[4578]: I1003 13:15:38.818524 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-dns-svc\") pod \"81676263-327d-4d19-8673-a44d5e576430\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " Oct 03 13:15:38 crc kubenswrapper[4578]: I1003 13:15:38.818544 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-ovsdbserver-sb\") pod \"81676263-327d-4d19-8673-a44d5e576430\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " Oct 03 13:15:38 crc kubenswrapper[4578]: I1003 13:15:38.818603 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g4sj\" (UniqueName: \"kubernetes.io/projected/81676263-327d-4d19-8673-a44d5e576430-kube-api-access-6g4sj\") pod \"81676263-327d-4d19-8673-a44d5e576430\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " Oct 03 13:15:38 crc kubenswrapper[4578]: I1003 13:15:38.818709 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-dns-swift-storage-0\") pod \"81676263-327d-4d19-8673-a44d5e576430\" (UID: \"81676263-327d-4d19-8673-a44d5e576430\") " Oct 03 13:15:38 crc kubenswrapper[4578]: I1003 13:15:38.850785 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81676263-327d-4d19-8673-a44d5e576430-kube-api-access-6g4sj" (OuterVolumeSpecName: "kube-api-access-6g4sj") pod "81676263-327d-4d19-8673-a44d5e576430" (UID: "81676263-327d-4d19-8673-a44d5e576430"). InnerVolumeSpecName "kube-api-access-6g4sj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:15:38 crc kubenswrapper[4578]: I1003 13:15:38.906592 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "81676263-327d-4d19-8673-a44d5e576430" (UID: "81676263-327d-4d19-8673-a44d5e576430"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:15:38 crc kubenswrapper[4578]: I1003 13:15:38.921061 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g4sj\" (UniqueName: \"kubernetes.io/projected/81676263-327d-4d19-8673-a44d5e576430-kube-api-access-6g4sj\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:38 crc kubenswrapper[4578]: I1003 13:15:38.921103 4578 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:38 crc kubenswrapper[4578]: I1003 13:15:38.930159 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "81676263-327d-4d19-8673-a44d5e576430" (UID: "81676263-327d-4d19-8673-a44d5e576430"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:15:38 crc kubenswrapper[4578]: I1003 13:15:38.961033 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "81676263-327d-4d19-8673-a44d5e576430" (UID: "81676263-327d-4d19-8673-a44d5e576430"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:15:38 crc kubenswrapper[4578]: I1003 13:15:38.978098 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "81676263-327d-4d19-8673-a44d5e576430" (UID: "81676263-327d-4d19-8673-a44d5e576430"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.008724 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-config" (OuterVolumeSpecName: "config") pod "81676263-327d-4d19-8673-a44d5e576430" (UID: "81676263-327d-4d19-8673-a44d5e576430"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.022823 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.022861 4578 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.022870 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.022882 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81676263-327d-4d19-8673-a44d5e576430-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.665445 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.665444 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-g7kjh" event={"ID":"81676263-327d-4d19-8673-a44d5e576430","Type":"ContainerDied","Data":"9296bd0a9e59d957fe83229af5326fb0d254487b9278fc15f157d9dc6c38a77c"} Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.666880 4578 scope.go:117] "RemoveContainer" containerID="4b1b34611cc3274a88742b81445a6c0ca5c68f7a7c316eceb9dad6764ab43967" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.682803 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a54925d7-6f48-4ce4-8062-c81a33abb815","Type":"ContainerStarted","Data":"d1bae298dfbc196fcc153db6f70177e81aaeb24969855544f5e065bf8bc80c0c"} Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.703885 4578 scope.go:117] "RemoveContainer" containerID="92c3b9ede7c0673b1000809f1f762eab7b156b40fca03f0eb4d352abae8eaaa1" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.707743 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-g7kjh"] Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.717657 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-g7kjh"] Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.743396 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jr79q"] Oct 03 13:15:39 crc kubenswrapper[4578]: E1003 13:15:39.743820 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81676263-327d-4d19-8673-a44d5e576430" containerName="init" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.743836 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="81676263-327d-4d19-8673-a44d5e576430" containerName="init" Oct 03 13:15:39 crc kubenswrapper[4578]: E1003 13:15:39.743856 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81676263-327d-4d19-8673-a44d5e576430" containerName="dnsmasq-dns" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.743862 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="81676263-327d-4d19-8673-a44d5e576430" containerName="dnsmasq-dns" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.744462 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="81676263-327d-4d19-8673-a44d5e576430" containerName="dnsmasq-dns" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.749838 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jr79q" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.755356 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jr79q"] Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.835579 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfc26ec2-4854-49fe-9c7b-5c548482c092-catalog-content\") pod \"certified-operators-jr79q\" (UID: \"dfc26ec2-4854-49fe-9c7b-5c548482c092\") " pod="openshift-marketplace/certified-operators-jr79q" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.835670 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mqrb\" (UniqueName: \"kubernetes.io/projected/dfc26ec2-4854-49fe-9c7b-5c548482c092-kube-api-access-5mqrb\") pod \"certified-operators-jr79q\" (UID: \"dfc26ec2-4854-49fe-9c7b-5c548482c092\") " pod="openshift-marketplace/certified-operators-jr79q" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.835840 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfc26ec2-4854-49fe-9c7b-5c548482c092-utilities\") pod \"certified-operators-jr79q\" (UID: \"dfc26ec2-4854-49fe-9c7b-5c548482c092\") " pod="openshift-marketplace/certified-operators-jr79q" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.937979 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfc26ec2-4854-49fe-9c7b-5c548482c092-utilities\") pod \"certified-operators-jr79q\" (UID: \"dfc26ec2-4854-49fe-9c7b-5c548482c092\") " pod="openshift-marketplace/certified-operators-jr79q" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.938122 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfc26ec2-4854-49fe-9c7b-5c548482c092-catalog-content\") pod \"certified-operators-jr79q\" (UID: \"dfc26ec2-4854-49fe-9c7b-5c548482c092\") " pod="openshift-marketplace/certified-operators-jr79q" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.938177 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mqrb\" (UniqueName: \"kubernetes.io/projected/dfc26ec2-4854-49fe-9c7b-5c548482c092-kube-api-access-5mqrb\") pod \"certified-operators-jr79q\" (UID: \"dfc26ec2-4854-49fe-9c7b-5c548482c092\") " pod="openshift-marketplace/certified-operators-jr79q" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.938459 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfc26ec2-4854-49fe-9c7b-5c548482c092-utilities\") pod \"certified-operators-jr79q\" (UID: \"dfc26ec2-4854-49fe-9c7b-5c548482c092\") " pod="openshift-marketplace/certified-operators-jr79q" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.938539 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfc26ec2-4854-49fe-9c7b-5c548482c092-catalog-content\") pod \"certified-operators-jr79q\" (UID: \"dfc26ec2-4854-49fe-9c7b-5c548482c092\") " pod="openshift-marketplace/certified-operators-jr79q" Oct 03 13:15:39 crc kubenswrapper[4578]: I1003 13:15:39.965611 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mqrb\" (UniqueName: \"kubernetes.io/projected/dfc26ec2-4854-49fe-9c7b-5c548482c092-kube-api-access-5mqrb\") pod \"certified-operators-jr79q\" (UID: \"dfc26ec2-4854-49fe-9c7b-5c548482c092\") " pod="openshift-marketplace/certified-operators-jr79q" Oct 03 13:15:40 crc kubenswrapper[4578]: I1003 13:15:40.089712 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jr79q" Oct 03 13:15:40 crc kubenswrapper[4578]: I1003 13:15:40.694573 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a54925d7-6f48-4ce4-8062-c81a33abb815","Type":"ContainerStarted","Data":"936e913fcbe354de4e348daab4d5f9685b00252c0a0a7bf93be543914c8d05bf"} Oct 03 13:15:40 crc kubenswrapper[4578]: I1003 13:15:40.696203 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 03 13:15:40 crc kubenswrapper[4578]: I1003 13:15:40.728902 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.298269617 podStartE2EDuration="5.728886589s" podCreationTimestamp="2025-10-03 13:15:35 +0000 UTC" firstStartedPulling="2025-10-03 13:15:36.533222964 +0000 UTC m=+1472.331695148" lastFinishedPulling="2025-10-03 13:15:39.963839936 +0000 UTC m=+1475.762312120" observedRunningTime="2025-10-03 13:15:40.724919225 +0000 UTC m=+1476.523391409" watchObservedRunningTime="2025-10-03 13:15:40.728886589 +0000 UTC m=+1476.527358773" Oct 03 13:15:40 crc kubenswrapper[4578]: I1003 13:15:40.753611 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jr79q"] Oct 03 13:15:40 crc kubenswrapper[4578]: I1003 13:15:40.921761 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81676263-327d-4d19-8673-a44d5e576430" path="/var/lib/kubelet/pods/81676263-327d-4d19-8673-a44d5e576430/volumes" Oct 03 13:15:41 crc kubenswrapper[4578]: I1003 13:15:41.704810 4578 generic.go:334] "Generic (PLEG): container finished" podID="dfc26ec2-4854-49fe-9c7b-5c548482c092" containerID="5545d3d880bbea8435e1ce43aad7f0a2d16f186af17ecc2bd65dc5fc2a672d26" exitCode=0 Oct 03 13:15:41 crc kubenswrapper[4578]: I1003 13:15:41.704879 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr79q" event={"ID":"dfc26ec2-4854-49fe-9c7b-5c548482c092","Type":"ContainerDied","Data":"5545d3d880bbea8435e1ce43aad7f0a2d16f186af17ecc2bd65dc5fc2a672d26"} Oct 03 13:15:41 crc kubenswrapper[4578]: I1003 13:15:41.705195 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr79q" event={"ID":"dfc26ec2-4854-49fe-9c7b-5c548482c092","Type":"ContainerStarted","Data":"ab6192501db15b837d443d0197beb5c8ff34606042fc1de2c1f76955208ddace"} Oct 03 13:15:42 crc kubenswrapper[4578]: I1003 13:15:42.717203 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr79q" event={"ID":"dfc26ec2-4854-49fe-9c7b-5c548482c092","Type":"ContainerStarted","Data":"5b6f0c812e1d7bb01b5d21e551f2d5c1e6f6aa4f6d52951afd385933aa21247a"} Oct 03 13:15:44 crc kubenswrapper[4578]: I1003 13:15:44.740091 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr79q" event={"ID":"dfc26ec2-4854-49fe-9c7b-5c548482c092","Type":"ContainerDied","Data":"5b6f0c812e1d7bb01b5d21e551f2d5c1e6f6aa4f6d52951afd385933aa21247a"} Oct 03 13:15:44 crc kubenswrapper[4578]: I1003 13:15:44.740042 4578 generic.go:334] "Generic (PLEG): container finished" podID="dfc26ec2-4854-49fe-9c7b-5c548482c092" containerID="5b6f0c812e1d7bb01b5d21e551f2d5c1e6f6aa4f6d52951afd385933aa21247a" exitCode=0 Oct 03 13:15:44 crc kubenswrapper[4578]: I1003 13:15:44.748709 4578 generic.go:334] "Generic (PLEG): container finished" podID="d3e10cb8-af89-4561-a4ae-804943cf9440" containerID="c4fd7d19b6ab5bea908d1b9c1dcde5509c71ba22e94c933a286e2563297c1ef8" exitCode=0 Oct 03 13:15:44 crc kubenswrapper[4578]: I1003 13:15:44.748758 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mz6wt" event={"ID":"d3e10cb8-af89-4561-a4ae-804943cf9440","Type":"ContainerDied","Data":"c4fd7d19b6ab5bea908d1b9c1dcde5509c71ba22e94c933a286e2563297c1ef8"} Oct 03 13:15:45 crc kubenswrapper[4578]: I1003 13:15:45.278460 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 13:15:45 crc kubenswrapper[4578]: I1003 13:15:45.278495 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 13:15:45 crc kubenswrapper[4578]: I1003 13:15:45.762190 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 03 13:15:45 crc kubenswrapper[4578]: I1003 13:15:45.764458 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr79q" event={"ID":"dfc26ec2-4854-49fe-9c7b-5c548482c092","Type":"ContainerStarted","Data":"b195ec676ca478886e5fd427c958ce41b1614a8a51667378b0ef4d9f531c04e1"} Oct 03 13:15:45 crc kubenswrapper[4578]: I1003 13:15:45.766139 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 03 13:15:45 crc kubenswrapper[4578]: I1003 13:15:45.773391 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 03 13:15:45 crc kubenswrapper[4578]: I1003 13:15:45.871189 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jr79q" podStartSLOduration=3.2672817690000002 podStartE2EDuration="6.871167434s" podCreationTimestamp="2025-10-03 13:15:39 +0000 UTC" firstStartedPulling="2025-10-03 13:15:41.706877253 +0000 UTC m=+1477.505349437" lastFinishedPulling="2025-10-03 13:15:45.310762918 +0000 UTC m=+1481.109235102" observedRunningTime="2025-10-03 13:15:45.818571602 +0000 UTC m=+1481.617043796" watchObservedRunningTime="2025-10-03 13:15:45.871167434 +0000 UTC m=+1481.669639608" Oct 03 13:15:46 crc kubenswrapper[4578]: I1003 13:15:46.234126 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mz6wt" Oct 03 13:15:46 crc kubenswrapper[4578]: I1003 13:15:46.325759 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="aa56cf15-10a3-4978-a7cd-1d0c32d1377c" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:15:46 crc kubenswrapper[4578]: I1003 13:15:46.325801 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="aa56cf15-10a3-4978-a7cd-1d0c32d1377c" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 13:15:46 crc kubenswrapper[4578]: I1003 13:15:46.371698 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3e10cb8-af89-4561-a4ae-804943cf9440-combined-ca-bundle\") pod \"d3e10cb8-af89-4561-a4ae-804943cf9440\" (UID: \"d3e10cb8-af89-4561-a4ae-804943cf9440\") " Oct 03 13:15:46 crc kubenswrapper[4578]: I1003 13:15:46.371740 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llp7r\" (UniqueName: \"kubernetes.io/projected/d3e10cb8-af89-4561-a4ae-804943cf9440-kube-api-access-llp7r\") pod \"d3e10cb8-af89-4561-a4ae-804943cf9440\" (UID: \"d3e10cb8-af89-4561-a4ae-804943cf9440\") " Oct 03 13:15:46 crc kubenswrapper[4578]: I1003 13:15:46.371775 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3e10cb8-af89-4561-a4ae-804943cf9440-config-data\") pod \"d3e10cb8-af89-4561-a4ae-804943cf9440\" (UID: \"d3e10cb8-af89-4561-a4ae-804943cf9440\") " Oct 03 13:15:46 crc kubenswrapper[4578]: I1003 13:15:46.371801 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3e10cb8-af89-4561-a4ae-804943cf9440-scripts\") pod \"d3e10cb8-af89-4561-a4ae-804943cf9440\" (UID: \"d3e10cb8-af89-4561-a4ae-804943cf9440\") " Oct 03 13:15:46 crc kubenswrapper[4578]: I1003 13:15:46.388694 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3e10cb8-af89-4561-a4ae-804943cf9440-kube-api-access-llp7r" (OuterVolumeSpecName: "kube-api-access-llp7r") pod "d3e10cb8-af89-4561-a4ae-804943cf9440" (UID: "d3e10cb8-af89-4561-a4ae-804943cf9440"). InnerVolumeSpecName "kube-api-access-llp7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:15:46 crc kubenswrapper[4578]: I1003 13:15:46.392149 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3e10cb8-af89-4561-a4ae-804943cf9440-scripts" (OuterVolumeSpecName: "scripts") pod "d3e10cb8-af89-4561-a4ae-804943cf9440" (UID: "d3e10cb8-af89-4561-a4ae-804943cf9440"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:46 crc kubenswrapper[4578]: I1003 13:15:46.420137 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3e10cb8-af89-4561-a4ae-804943cf9440-config-data" (OuterVolumeSpecName: "config-data") pod "d3e10cb8-af89-4561-a4ae-804943cf9440" (UID: "d3e10cb8-af89-4561-a4ae-804943cf9440"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:46 crc kubenswrapper[4578]: I1003 13:15:46.453079 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3e10cb8-af89-4561-a4ae-804943cf9440-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3e10cb8-af89-4561-a4ae-804943cf9440" (UID: "d3e10cb8-af89-4561-a4ae-804943cf9440"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:46 crc kubenswrapper[4578]: I1003 13:15:46.480869 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3e10cb8-af89-4561-a4ae-804943cf9440-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:46 crc kubenswrapper[4578]: I1003 13:15:46.481092 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llp7r\" (UniqueName: \"kubernetes.io/projected/d3e10cb8-af89-4561-a4ae-804943cf9440-kube-api-access-llp7r\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:46 crc kubenswrapper[4578]: I1003 13:15:46.481102 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3e10cb8-af89-4561-a4ae-804943cf9440-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:46 crc kubenswrapper[4578]: I1003 13:15:46.481111 4578 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3e10cb8-af89-4561-a4ae-804943cf9440-scripts\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:46 crc kubenswrapper[4578]: I1003 13:15:46.777000 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mz6wt" Oct 03 13:15:46 crc kubenswrapper[4578]: I1003 13:15:46.777607 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mz6wt" event={"ID":"d3e10cb8-af89-4561-a4ae-804943cf9440","Type":"ContainerDied","Data":"e0a4e8584b26ffe81f3dc9bc5a49109bbf05a323a946355cc3e43b56e2c9ad4d"} Oct 03 13:15:46 crc kubenswrapper[4578]: I1003 13:15:46.777658 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0a4e8584b26ffe81f3dc9bc5a49109bbf05a323a946355cc3e43b56e2c9ad4d" Oct 03 13:15:46 crc kubenswrapper[4578]: I1003 13:15:46.815406 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 03 13:15:47 crc kubenswrapper[4578]: I1003 13:15:47.042559 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:15:47 crc kubenswrapper[4578]: I1003 13:15:47.043450 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="aa56cf15-10a3-4978-a7cd-1d0c32d1377c" containerName="nova-api-log" containerID="cri-o://f9ee50c58c5c748dd8c43e06176ea153e9a4e6056f78a3923f9a4c2d9b9fe25d" gracePeriod=30 Oct 03 13:15:47 crc kubenswrapper[4578]: I1003 13:15:47.044219 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="aa56cf15-10a3-4978-a7cd-1d0c32d1377c" containerName="nova-api-api" containerID="cri-o://afaed3a1fb21a22bb87ac250d1cbd6ff3af9bbe14f49dc554001442d1544a936" gracePeriod=30 Oct 03 13:15:47 crc kubenswrapper[4578]: I1003 13:15:47.120129 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:15:47 crc kubenswrapper[4578]: I1003 13:15:47.120357 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="9de81a4c-5730-443b-9a48-1923ddb3c139" containerName="nova-scheduler-scheduler" containerID="cri-o://84b5a9682c1a09a7c23497c8aa1c6992b406a591d51024911638716e9cca16a5" gracePeriod=30 Oct 03 13:15:47 crc kubenswrapper[4578]: I1003 13:15:47.150784 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:15:47 crc kubenswrapper[4578]: I1003 13:15:47.788243 4578 generic.go:334] "Generic (PLEG): container finished" podID="aa56cf15-10a3-4978-a7cd-1d0c32d1377c" containerID="f9ee50c58c5c748dd8c43e06176ea153e9a4e6056f78a3923f9a4c2d9b9fe25d" exitCode=143 Oct 03 13:15:47 crc kubenswrapper[4578]: I1003 13:15:47.788312 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aa56cf15-10a3-4978-a7cd-1d0c32d1377c","Type":"ContainerDied","Data":"f9ee50c58c5c748dd8c43e06176ea153e9a4e6056f78a3923f9a4c2d9b9fe25d"} Oct 03 13:15:48 crc kubenswrapper[4578]: I1003 13:15:48.796931 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="34b83c0e-8c44-4ceb-abae-35551eb3f183" containerName="nova-metadata-log" containerID="cri-o://4b057177e19710c3252908b4c28cc2e673d957c473c87d7dc0e2c3c5305d9647" gracePeriod=30 Oct 03 13:15:48 crc kubenswrapper[4578]: I1003 13:15:48.797007 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="34b83c0e-8c44-4ceb-abae-35551eb3f183" containerName="nova-metadata-metadata" containerID="cri-o://e1cbf44efb40d92f0202b86942cdf00f47b5fbd2338ecf5dcd010d207d07a59f" gracePeriod=30 Oct 03 13:15:49 crc kubenswrapper[4578]: I1003 13:15:49.808198 4578 generic.go:334] "Generic (PLEG): container finished" podID="34b83c0e-8c44-4ceb-abae-35551eb3f183" containerID="4b057177e19710c3252908b4c28cc2e673d957c473c87d7dc0e2c3c5305d9647" exitCode=143 Oct 03 13:15:49 crc kubenswrapper[4578]: I1003 13:15:49.808308 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"34b83c0e-8c44-4ceb-abae-35551eb3f183","Type":"ContainerDied","Data":"4b057177e19710c3252908b4c28cc2e673d957c473c87d7dc0e2c3c5305d9647"} Oct 03 13:15:50 crc kubenswrapper[4578]: I1003 13:15:50.090852 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jr79q" Oct 03 13:15:50 crc kubenswrapper[4578]: I1003 13:15:50.090900 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jr79q" Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.169420 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-jr79q" podUID="dfc26ec2-4854-49fe-9c7b-5c548482c092" containerName="registry-server" probeResult="failure" output=< Oct 03 13:15:51 crc kubenswrapper[4578]: timeout: failed to connect service ":50051" within 1s Oct 03 13:15:51 crc kubenswrapper[4578]: > Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.378772 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.507898 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9de81a4c-5730-443b-9a48-1923ddb3c139-config-data\") pod \"9de81a4c-5730-443b-9a48-1923ddb3c139\" (UID: \"9de81a4c-5730-443b-9a48-1923ddb3c139\") " Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.508300 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5qt6\" (UniqueName: \"kubernetes.io/projected/9de81a4c-5730-443b-9a48-1923ddb3c139-kube-api-access-r5qt6\") pod \"9de81a4c-5730-443b-9a48-1923ddb3c139\" (UID: \"9de81a4c-5730-443b-9a48-1923ddb3c139\") " Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.508438 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9de81a4c-5730-443b-9a48-1923ddb3c139-combined-ca-bundle\") pod \"9de81a4c-5730-443b-9a48-1923ddb3c139\" (UID: \"9de81a4c-5730-443b-9a48-1923ddb3c139\") " Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.532939 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9de81a4c-5730-443b-9a48-1923ddb3c139-kube-api-access-r5qt6" (OuterVolumeSpecName: "kube-api-access-r5qt6") pod "9de81a4c-5730-443b-9a48-1923ddb3c139" (UID: "9de81a4c-5730-443b-9a48-1923ddb3c139"). InnerVolumeSpecName "kube-api-access-r5qt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.544295 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9de81a4c-5730-443b-9a48-1923ddb3c139-config-data" (OuterVolumeSpecName: "config-data") pod "9de81a4c-5730-443b-9a48-1923ddb3c139" (UID: "9de81a4c-5730-443b-9a48-1923ddb3c139"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.558699 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9de81a4c-5730-443b-9a48-1923ddb3c139-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9de81a4c-5730-443b-9a48-1923ddb3c139" (UID: "9de81a4c-5730-443b-9a48-1923ddb3c139"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.611274 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9de81a4c-5730-443b-9a48-1923ddb3c139-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.611309 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5qt6\" (UniqueName: \"kubernetes.io/projected/9de81a4c-5730-443b-9a48-1923ddb3c139-kube-api-access-r5qt6\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.611319 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9de81a4c-5730-443b-9a48-1923ddb3c139-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.828874 4578 generic.go:334] "Generic (PLEG): container finished" podID="9de81a4c-5730-443b-9a48-1923ddb3c139" containerID="84b5a9682c1a09a7c23497c8aa1c6992b406a591d51024911638716e9cca16a5" exitCode=0 Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.829057 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9de81a4c-5730-443b-9a48-1923ddb3c139","Type":"ContainerDied","Data":"84b5a9682c1a09a7c23497c8aa1c6992b406a591d51024911638716e9cca16a5"} Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.829187 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"9de81a4c-5730-443b-9a48-1923ddb3c139","Type":"ContainerDied","Data":"5c4160ee31028df1e84536d44f54f1e4fb054b1da3e286e7e1ff849c1246231c"} Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.829257 4578 scope.go:117] "RemoveContainer" containerID="84b5a9682c1a09a7c23497c8aa1c6992b406a591d51024911638716e9cca16a5" Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.829135 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.871803 4578 scope.go:117] "RemoveContainer" containerID="84b5a9682c1a09a7c23497c8aa1c6992b406a591d51024911638716e9cca16a5" Oct 03 13:15:51 crc kubenswrapper[4578]: E1003 13:15:51.872212 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84b5a9682c1a09a7c23497c8aa1c6992b406a591d51024911638716e9cca16a5\": container with ID starting with 84b5a9682c1a09a7c23497c8aa1c6992b406a591d51024911638716e9cca16a5 not found: ID does not exist" containerID="84b5a9682c1a09a7c23497c8aa1c6992b406a591d51024911638716e9cca16a5" Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.872355 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84b5a9682c1a09a7c23497c8aa1c6992b406a591d51024911638716e9cca16a5"} err="failed to get container status \"84b5a9682c1a09a7c23497c8aa1c6992b406a591d51024911638716e9cca16a5\": rpc error: code = NotFound desc = could not find container \"84b5a9682c1a09a7c23497c8aa1c6992b406a591d51024911638716e9cca16a5\": container with ID starting with 84b5a9682c1a09a7c23497c8aa1c6992b406a591d51024911638716e9cca16a5 not found: ID does not exist" Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.872800 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.885741 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.937192 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:15:51 crc kubenswrapper[4578]: E1003 13:15:51.937580 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3e10cb8-af89-4561-a4ae-804943cf9440" containerName="nova-manage" Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.937595 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3e10cb8-af89-4561-a4ae-804943cf9440" containerName="nova-manage" Oct 03 13:15:51 crc kubenswrapper[4578]: E1003 13:15:51.937618 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9de81a4c-5730-443b-9a48-1923ddb3c139" containerName="nova-scheduler-scheduler" Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.937624 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="9de81a4c-5730-443b-9a48-1923ddb3c139" containerName="nova-scheduler-scheduler" Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.937831 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3e10cb8-af89-4561-a4ae-804943cf9440" containerName="nova-manage" Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.937860 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="9de81a4c-5730-443b-9a48-1923ddb3c139" containerName="nova-scheduler-scheduler" Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.938485 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.944800 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 03 13:15:51 crc kubenswrapper[4578]: I1003 13:15:51.960236 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.031173 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b3f582b-6953-464d-902b-f008def7a0fe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5b3f582b-6953-464d-902b-f008def7a0fe\") " pod="openstack/nova-scheduler-0" Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.031229 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8gkf\" (UniqueName: \"kubernetes.io/projected/5b3f582b-6953-464d-902b-f008def7a0fe-kube-api-access-b8gkf\") pod \"nova-scheduler-0\" (UID: \"5b3f582b-6953-464d-902b-f008def7a0fe\") " pod="openstack/nova-scheduler-0" Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.031330 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b3f582b-6953-464d-902b-f008def7a0fe-config-data\") pod \"nova-scheduler-0\" (UID: \"5b3f582b-6953-464d-902b-f008def7a0fe\") " pod="openstack/nova-scheduler-0" Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.133211 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b3f582b-6953-464d-902b-f008def7a0fe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5b3f582b-6953-464d-902b-f008def7a0fe\") " pod="openstack/nova-scheduler-0" Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.134021 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8gkf\" (UniqueName: \"kubernetes.io/projected/5b3f582b-6953-464d-902b-f008def7a0fe-kube-api-access-b8gkf\") pod \"nova-scheduler-0\" (UID: \"5b3f582b-6953-464d-902b-f008def7a0fe\") " pod="openstack/nova-scheduler-0" Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.134175 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b3f582b-6953-464d-902b-f008def7a0fe-config-data\") pod \"nova-scheduler-0\" (UID: \"5b3f582b-6953-464d-902b-f008def7a0fe\") " pod="openstack/nova-scheduler-0" Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.137107 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b3f582b-6953-464d-902b-f008def7a0fe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5b3f582b-6953-464d-902b-f008def7a0fe\") " pod="openstack/nova-scheduler-0" Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.137954 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b3f582b-6953-464d-902b-f008def7a0fe-config-data\") pod \"nova-scheduler-0\" (UID: \"5b3f582b-6953-464d-902b-f008def7a0fe\") " pod="openstack/nova-scheduler-0" Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.155107 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8gkf\" (UniqueName: \"kubernetes.io/projected/5b3f582b-6953-464d-902b-f008def7a0fe-kube-api-access-b8gkf\") pod \"nova-scheduler-0\" (UID: \"5b3f582b-6953-464d-902b-f008def7a0fe\") " pod="openstack/nova-scheduler-0" Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.259606 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="34b83c0e-8c44-4ceb-abae-35551eb3f183" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": read tcp 10.217.0.2:33394->10.217.0.197:8775: read: connection reset by peer" Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.265062 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="34b83c0e-8c44-4ceb-abae-35551eb3f183" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": read tcp 10.217.0.2:33390->10.217.0.197:8775: read: connection reset by peer" Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.265708 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.775517 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 03 13:15:52 crc kubenswrapper[4578]: W1003 13:15:52.824495 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b3f582b_6953_464d_902b_f008def7a0fe.slice/crio-fd7151e1ff27d553a256be44ff449b87f705560988d1b92475fd7ed141848aa6 WatchSource:0}: Error finding container fd7151e1ff27d553a256be44ff449b87f705560988d1b92475fd7ed141848aa6: Status 404 returned error can't find the container with id fd7151e1ff27d553a256be44ff449b87f705560988d1b92475fd7ed141848aa6 Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.857817 4578 generic.go:334] "Generic (PLEG): container finished" podID="aa56cf15-10a3-4978-a7cd-1d0c32d1377c" containerID="afaed3a1fb21a22bb87ac250d1cbd6ff3af9bbe14f49dc554001442d1544a936" exitCode=0 Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.857875 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aa56cf15-10a3-4978-a7cd-1d0c32d1377c","Type":"ContainerDied","Data":"afaed3a1fb21a22bb87ac250d1cbd6ff3af9bbe14f49dc554001442d1544a936"} Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.859584 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5b3f582b-6953-464d-902b-f008def7a0fe","Type":"ContainerStarted","Data":"fd7151e1ff27d553a256be44ff449b87f705560988d1b92475fd7ed141848aa6"} Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.880924 4578 generic.go:334] "Generic (PLEG): container finished" podID="34b83c0e-8c44-4ceb-abae-35551eb3f183" containerID="e1cbf44efb40d92f0202b86942cdf00f47b5fbd2338ecf5dcd010d207d07a59f" exitCode=0 Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.880991 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"34b83c0e-8c44-4ceb-abae-35551eb3f183","Type":"ContainerDied","Data":"e1cbf44efb40d92f0202b86942cdf00f47b5fbd2338ecf5dcd010d207d07a59f"} Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.881022 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"34b83c0e-8c44-4ceb-abae-35551eb3f183","Type":"ContainerDied","Data":"da4a0e9f1b027c0e7415a9c4d500f0eaddc681a58f73c0316c17f5ad0cf9b7f8"} Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.881034 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da4a0e9f1b027c0e7415a9c4d500f0eaddc681a58f73c0316c17f5ad0cf9b7f8" Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.881740 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.925329 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9de81a4c-5730-443b-9a48-1923ddb3c139" path="/var/lib/kubelet/pods/9de81a4c-5730-443b-9a48-1923ddb3c139/volumes" Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.955246 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/34b83c0e-8c44-4ceb-abae-35551eb3f183-nova-metadata-tls-certs\") pod \"34b83c0e-8c44-4ceb-abae-35551eb3f183\" (UID: \"34b83c0e-8c44-4ceb-abae-35551eb3f183\") " Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.955299 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cldsw\" (UniqueName: \"kubernetes.io/projected/34b83c0e-8c44-4ceb-abae-35551eb3f183-kube-api-access-cldsw\") pod \"34b83c0e-8c44-4ceb-abae-35551eb3f183\" (UID: \"34b83c0e-8c44-4ceb-abae-35551eb3f183\") " Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.955322 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34b83c0e-8c44-4ceb-abae-35551eb3f183-combined-ca-bundle\") pod \"34b83c0e-8c44-4ceb-abae-35551eb3f183\" (UID: \"34b83c0e-8c44-4ceb-abae-35551eb3f183\") " Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.955469 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34b83c0e-8c44-4ceb-abae-35551eb3f183-logs\") pod \"34b83c0e-8c44-4ceb-abae-35551eb3f183\" (UID: \"34b83c0e-8c44-4ceb-abae-35551eb3f183\") " Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.955551 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34b83c0e-8c44-4ceb-abae-35551eb3f183-config-data\") pod \"34b83c0e-8c44-4ceb-abae-35551eb3f183\" (UID: \"34b83c0e-8c44-4ceb-abae-35551eb3f183\") " Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.976002 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34b83c0e-8c44-4ceb-abae-35551eb3f183-logs" (OuterVolumeSpecName: "logs") pod "34b83c0e-8c44-4ceb-abae-35551eb3f183" (UID: "34b83c0e-8c44-4ceb-abae-35551eb3f183"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:15:52 crc kubenswrapper[4578]: I1003 13:15:52.993924 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34b83c0e-8c44-4ceb-abae-35551eb3f183-kube-api-access-cldsw" (OuterVolumeSpecName: "kube-api-access-cldsw") pod "34b83c0e-8c44-4ceb-abae-35551eb3f183" (UID: "34b83c0e-8c44-4ceb-abae-35551eb3f183"). InnerVolumeSpecName "kube-api-access-cldsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.044986 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34b83c0e-8c44-4ceb-abae-35551eb3f183-config-data" (OuterVolumeSpecName: "config-data") pod "34b83c0e-8c44-4ceb-abae-35551eb3f183" (UID: "34b83c0e-8c44-4ceb-abae-35551eb3f183"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.059932 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cldsw\" (UniqueName: \"kubernetes.io/projected/34b83c0e-8c44-4ceb-abae-35551eb3f183-kube-api-access-cldsw\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.059956 4578 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34b83c0e-8c44-4ceb-abae-35551eb3f183-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.059965 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34b83c0e-8c44-4ceb-abae-35551eb3f183-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.069813 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.089104 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34b83c0e-8c44-4ceb-abae-35551eb3f183-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34b83c0e-8c44-4ceb-abae-35551eb3f183" (UID: "34b83c0e-8c44-4ceb-abae-35551eb3f183"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.118568 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34b83c0e-8c44-4ceb-abae-35551eb3f183-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "34b83c0e-8c44-4ceb-abae-35551eb3f183" (UID: "34b83c0e-8c44-4ceb-abae-35551eb3f183"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.161355 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-config-data\") pod \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.161717 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-internal-tls-certs\") pod \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.161846 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-logs\") pod \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.161890 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-combined-ca-bundle\") pod \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.161919 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-public-tls-certs\") pod \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.161952 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vqdp\" (UniqueName: \"kubernetes.io/projected/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-kube-api-access-6vqdp\") pod \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\" (UID: \"aa56cf15-10a3-4978-a7cd-1d0c32d1377c\") " Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.162728 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-logs" (OuterVolumeSpecName: "logs") pod "aa56cf15-10a3-4978-a7cd-1d0c32d1377c" (UID: "aa56cf15-10a3-4978-a7cd-1d0c32d1377c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.163051 4578 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-logs\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.163066 4578 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/34b83c0e-8c44-4ceb-abae-35551eb3f183-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.163076 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34b83c0e-8c44-4ceb-abae-35551eb3f183-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.182322 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-kube-api-access-6vqdp" (OuterVolumeSpecName: "kube-api-access-6vqdp") pod "aa56cf15-10a3-4978-a7cd-1d0c32d1377c" (UID: "aa56cf15-10a3-4978-a7cd-1d0c32d1377c"). InnerVolumeSpecName "kube-api-access-6vqdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.253739 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-config-data" (OuterVolumeSpecName: "config-data") pod "aa56cf15-10a3-4978-a7cd-1d0c32d1377c" (UID: "aa56cf15-10a3-4978-a7cd-1d0c32d1377c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.258811 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa56cf15-10a3-4978-a7cd-1d0c32d1377c" (UID: "aa56cf15-10a3-4978-a7cd-1d0c32d1377c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.265122 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.265152 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vqdp\" (UniqueName: \"kubernetes.io/projected/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-kube-api-access-6vqdp\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.265165 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.291489 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "aa56cf15-10a3-4978-a7cd-1d0c32d1377c" (UID: "aa56cf15-10a3-4978-a7cd-1d0c32d1377c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.357748 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "aa56cf15-10a3-4978-a7cd-1d0c32d1377c" (UID: "aa56cf15-10a3-4978-a7cd-1d0c32d1377c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.369849 4578 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.369879 4578 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa56cf15-10a3-4978-a7cd-1d0c32d1377c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.893956 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"aa56cf15-10a3-4978-a7cd-1d0c32d1377c","Type":"ContainerDied","Data":"cf0c720e06b7cd264814a0a477a88a4be612b3c3a32536da0b8dd3b3c7ea5a8d"} Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.894025 4578 scope.go:117] "RemoveContainer" containerID="afaed3a1fb21a22bb87ac250d1cbd6ff3af9bbe14f49dc554001442d1544a936" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.894065 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.897285 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5b3f582b-6953-464d-902b-f008def7a0fe","Type":"ContainerStarted","Data":"c00822e34a1a6363ab35f5f921190514e882dcd1dfc295c820af5e791f401a55"} Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.897336 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.923472 4578 scope.go:117] "RemoveContainer" containerID="f9ee50c58c5c748dd8c43e06176ea153e9a4e6056f78a3923f9a4c2d9b9fe25d" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.938810 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.938790327 podStartE2EDuration="2.938790327s" podCreationTimestamp="2025-10-03 13:15:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:15:53.933872542 +0000 UTC m=+1489.732344726" watchObservedRunningTime="2025-10-03 13:15:53.938790327 +0000 UTC m=+1489.737262511" Oct 03 13:15:53 crc kubenswrapper[4578]: I1003 13:15:53.970798 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.002114 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.041300 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.068913 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.083368 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 03 13:15:54 crc kubenswrapper[4578]: E1003 13:15:54.084991 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34b83c0e-8c44-4ceb-abae-35551eb3f183" containerName="nova-metadata-metadata" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.085014 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="34b83c0e-8c44-4ceb-abae-35551eb3f183" containerName="nova-metadata-metadata" Oct 03 13:15:54 crc kubenswrapper[4578]: E1003 13:15:54.085065 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa56cf15-10a3-4978-a7cd-1d0c32d1377c" containerName="nova-api-log" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.085073 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa56cf15-10a3-4978-a7cd-1d0c32d1377c" containerName="nova-api-log" Oct 03 13:15:54 crc kubenswrapper[4578]: E1003 13:15:54.085107 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa56cf15-10a3-4978-a7cd-1d0c32d1377c" containerName="nova-api-api" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.085179 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa56cf15-10a3-4978-a7cd-1d0c32d1377c" containerName="nova-api-api" Oct 03 13:15:54 crc kubenswrapper[4578]: E1003 13:15:54.085253 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34b83c0e-8c44-4ceb-abae-35551eb3f183" containerName="nova-metadata-log" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.085267 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="34b83c0e-8c44-4ceb-abae-35551eb3f183" containerName="nova-metadata-log" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.085928 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa56cf15-10a3-4978-a7cd-1d0c32d1377c" containerName="nova-api-log" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.085957 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="34b83c0e-8c44-4ceb-abae-35551eb3f183" containerName="nova-metadata-log" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.085990 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa56cf15-10a3-4978-a7cd-1d0c32d1377c" containerName="nova-api-api" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.086009 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="34b83c0e-8c44-4ceb-abae-35551eb3f183" containerName="nova-metadata-metadata" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.089808 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.093482 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.094396 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.095430 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.101520 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.101810 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.104149 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.104347 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.112029 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.121348 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.190891 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89f4ced5-a0c4-4b41-9c87-64377d0fc4bd-logs\") pod \"nova-metadata-0\" (UID: \"89f4ced5-a0c4-4b41-9c87-64377d0fc4bd\") " pod="openstack/nova-metadata-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.191235 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/25b861ed-9dfe-4494-96a0-03fa994f1775-public-tls-certs\") pod \"nova-api-0\" (UID: \"25b861ed-9dfe-4494-96a0-03fa994f1775\") " pod="openstack/nova-api-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.191347 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/25b861ed-9dfe-4494-96a0-03fa994f1775-internal-tls-certs\") pod \"nova-api-0\" (UID: \"25b861ed-9dfe-4494-96a0-03fa994f1775\") " pod="openstack/nova-api-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.191478 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25b861ed-9dfe-4494-96a0-03fa994f1775-config-data\") pod \"nova-api-0\" (UID: \"25b861ed-9dfe-4494-96a0-03fa994f1775\") " pod="openstack/nova-api-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.191553 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/89f4ced5-a0c4-4b41-9c87-64377d0fc4bd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"89f4ced5-a0c4-4b41-9c87-64377d0fc4bd\") " pod="openstack/nova-metadata-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.191643 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25b861ed-9dfe-4494-96a0-03fa994f1775-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"25b861ed-9dfe-4494-96a0-03fa994f1775\") " pod="openstack/nova-api-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.191832 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vs4ds\" (UniqueName: \"kubernetes.io/projected/25b861ed-9dfe-4494-96a0-03fa994f1775-kube-api-access-vs4ds\") pod \"nova-api-0\" (UID: \"25b861ed-9dfe-4494-96a0-03fa994f1775\") " pod="openstack/nova-api-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.192003 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89f4ced5-a0c4-4b41-9c87-64377d0fc4bd-config-data\") pod \"nova-metadata-0\" (UID: \"89f4ced5-a0c4-4b41-9c87-64377d0fc4bd\") " pod="openstack/nova-metadata-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.192045 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25b861ed-9dfe-4494-96a0-03fa994f1775-logs\") pod \"nova-api-0\" (UID: \"25b861ed-9dfe-4494-96a0-03fa994f1775\") " pod="openstack/nova-api-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.192246 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89f4ced5-a0c4-4b41-9c87-64377d0fc4bd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"89f4ced5-a0c4-4b41-9c87-64377d0fc4bd\") " pod="openstack/nova-metadata-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.192312 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbpgg\" (UniqueName: \"kubernetes.io/projected/89f4ced5-a0c4-4b41-9c87-64377d0fc4bd-kube-api-access-hbpgg\") pod \"nova-metadata-0\" (UID: \"89f4ced5-a0c4-4b41-9c87-64377d0fc4bd\") " pod="openstack/nova-metadata-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.294348 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89f4ced5-a0c4-4b41-9c87-64377d0fc4bd-config-data\") pod \"nova-metadata-0\" (UID: \"89f4ced5-a0c4-4b41-9c87-64377d0fc4bd\") " pod="openstack/nova-metadata-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.294394 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25b861ed-9dfe-4494-96a0-03fa994f1775-logs\") pod \"nova-api-0\" (UID: \"25b861ed-9dfe-4494-96a0-03fa994f1775\") " pod="openstack/nova-api-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.294439 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89f4ced5-a0c4-4b41-9c87-64377d0fc4bd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"89f4ced5-a0c4-4b41-9c87-64377d0fc4bd\") " pod="openstack/nova-metadata-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.294463 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbpgg\" (UniqueName: \"kubernetes.io/projected/89f4ced5-a0c4-4b41-9c87-64377d0fc4bd-kube-api-access-hbpgg\") pod \"nova-metadata-0\" (UID: \"89f4ced5-a0c4-4b41-9c87-64377d0fc4bd\") " pod="openstack/nova-metadata-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.294506 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89f4ced5-a0c4-4b41-9c87-64377d0fc4bd-logs\") pod \"nova-metadata-0\" (UID: \"89f4ced5-a0c4-4b41-9c87-64377d0fc4bd\") " pod="openstack/nova-metadata-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.294532 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/25b861ed-9dfe-4494-96a0-03fa994f1775-public-tls-certs\") pod \"nova-api-0\" (UID: \"25b861ed-9dfe-4494-96a0-03fa994f1775\") " pod="openstack/nova-api-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.294558 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/25b861ed-9dfe-4494-96a0-03fa994f1775-internal-tls-certs\") pod \"nova-api-0\" (UID: \"25b861ed-9dfe-4494-96a0-03fa994f1775\") " pod="openstack/nova-api-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.294604 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25b861ed-9dfe-4494-96a0-03fa994f1775-config-data\") pod \"nova-api-0\" (UID: \"25b861ed-9dfe-4494-96a0-03fa994f1775\") " pod="openstack/nova-api-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.294620 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/89f4ced5-a0c4-4b41-9c87-64377d0fc4bd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"89f4ced5-a0c4-4b41-9c87-64377d0fc4bd\") " pod="openstack/nova-metadata-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.294655 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25b861ed-9dfe-4494-96a0-03fa994f1775-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"25b861ed-9dfe-4494-96a0-03fa994f1775\") " pod="openstack/nova-api-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.294676 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vs4ds\" (UniqueName: \"kubernetes.io/projected/25b861ed-9dfe-4494-96a0-03fa994f1775-kube-api-access-vs4ds\") pod \"nova-api-0\" (UID: \"25b861ed-9dfe-4494-96a0-03fa994f1775\") " pod="openstack/nova-api-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.295357 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25b861ed-9dfe-4494-96a0-03fa994f1775-logs\") pod \"nova-api-0\" (UID: \"25b861ed-9dfe-4494-96a0-03fa994f1775\") " pod="openstack/nova-api-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.296082 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89f4ced5-a0c4-4b41-9c87-64377d0fc4bd-logs\") pod \"nova-metadata-0\" (UID: \"89f4ced5-a0c4-4b41-9c87-64377d0fc4bd\") " pod="openstack/nova-metadata-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.317727 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/25b861ed-9dfe-4494-96a0-03fa994f1775-public-tls-certs\") pod \"nova-api-0\" (UID: \"25b861ed-9dfe-4494-96a0-03fa994f1775\") " pod="openstack/nova-api-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.317941 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89f4ced5-a0c4-4b41-9c87-64377d0fc4bd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"89f4ced5-a0c4-4b41-9c87-64377d0fc4bd\") " pod="openstack/nova-metadata-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.318146 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89f4ced5-a0c4-4b41-9c87-64377d0fc4bd-config-data\") pod \"nova-metadata-0\" (UID: \"89f4ced5-a0c4-4b41-9c87-64377d0fc4bd\") " pod="openstack/nova-metadata-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.318939 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/89f4ced5-a0c4-4b41-9c87-64377d0fc4bd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"89f4ced5-a0c4-4b41-9c87-64377d0fc4bd\") " pod="openstack/nova-metadata-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.319049 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25b861ed-9dfe-4494-96a0-03fa994f1775-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"25b861ed-9dfe-4494-96a0-03fa994f1775\") " pod="openstack/nova-api-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.319277 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25b861ed-9dfe-4494-96a0-03fa994f1775-config-data\") pod \"nova-api-0\" (UID: \"25b861ed-9dfe-4494-96a0-03fa994f1775\") " pod="openstack/nova-api-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.322687 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbpgg\" (UniqueName: \"kubernetes.io/projected/89f4ced5-a0c4-4b41-9c87-64377d0fc4bd-kube-api-access-hbpgg\") pod \"nova-metadata-0\" (UID: \"89f4ced5-a0c4-4b41-9c87-64377d0fc4bd\") " pod="openstack/nova-metadata-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.322705 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vs4ds\" (UniqueName: \"kubernetes.io/projected/25b861ed-9dfe-4494-96a0-03fa994f1775-kube-api-access-vs4ds\") pod \"nova-api-0\" (UID: \"25b861ed-9dfe-4494-96a0-03fa994f1775\") " pod="openstack/nova-api-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.322708 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/25b861ed-9dfe-4494-96a0-03fa994f1775-internal-tls-certs\") pod \"nova-api-0\" (UID: \"25b861ed-9dfe-4494-96a0-03fa994f1775\") " pod="openstack/nova-api-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.427297 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.438814 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.928140 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34b83c0e-8c44-4ceb-abae-35551eb3f183" path="/var/lib/kubelet/pods/34b83c0e-8c44-4ceb-abae-35551eb3f183/volumes" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.932511 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa56cf15-10a3-4978-a7cd-1d0c32d1377c" path="/var/lib/kubelet/pods/aa56cf15-10a3-4978-a7cd-1d0c32d1377c/volumes" Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.988675 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 03 13:15:54 crc kubenswrapper[4578]: I1003 13:15:54.999927 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 03 13:15:55 crc kubenswrapper[4578]: I1003 13:15:55.920301 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"25b861ed-9dfe-4494-96a0-03fa994f1775","Type":"ContainerStarted","Data":"fa433966eeab11c3790a545a1f9bd084441421c3bf2095aadb911e8c35269553"} Oct 03 13:15:55 crc kubenswrapper[4578]: I1003 13:15:55.920869 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"25b861ed-9dfe-4494-96a0-03fa994f1775","Type":"ContainerStarted","Data":"4ddae4117bf13256c5eb46e2f1efe78b0e08d71faea6a237a94d41501171ccfd"} Oct 03 13:15:55 crc kubenswrapper[4578]: I1003 13:15:55.920888 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"25b861ed-9dfe-4494-96a0-03fa994f1775","Type":"ContainerStarted","Data":"0c3051f69180a4582ce1dc7509b25f32dafadda7ecab7f9f0b300e38bbda5158"} Oct 03 13:15:55 crc kubenswrapper[4578]: I1003 13:15:55.934277 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"89f4ced5-a0c4-4b41-9c87-64377d0fc4bd","Type":"ContainerStarted","Data":"bf5b56a7401df828cc057bfcb9e52a6e121fc2b0e17b85b4f801881ca3e63969"} Oct 03 13:15:55 crc kubenswrapper[4578]: I1003 13:15:55.934329 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"89f4ced5-a0c4-4b41-9c87-64377d0fc4bd","Type":"ContainerStarted","Data":"b531f9c7ee3558c3445cbd53a7092e7c7063fb383fd0f50b561b5fbe4b0905e3"} Oct 03 13:15:55 crc kubenswrapper[4578]: I1003 13:15:55.934342 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"89f4ced5-a0c4-4b41-9c87-64377d0fc4bd","Type":"ContainerStarted","Data":"a1d92ab29f9fcde9eb6888c24edaff33b6c4485acd5ad0ba2fa70439a98a2b03"} Oct 03 13:15:55 crc kubenswrapper[4578]: I1003 13:15:55.975449 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.975432048 podStartE2EDuration="2.975432048s" podCreationTimestamp="2025-10-03 13:15:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:15:55.969007176 +0000 UTC m=+1491.767479370" watchObservedRunningTime="2025-10-03 13:15:55.975432048 +0000 UTC m=+1491.773904232" Oct 03 13:15:57 crc kubenswrapper[4578]: I1003 13:15:57.266423 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 03 13:15:59 crc kubenswrapper[4578]: I1003 13:15:59.440468 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 13:15:59 crc kubenswrapper[4578]: I1003 13:15:59.441464 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 03 13:16:00 crc kubenswrapper[4578]: I1003 13:16:00.172802 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jr79q" Oct 03 13:16:00 crc kubenswrapper[4578]: I1003 13:16:00.199762 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=7.199695392 podStartE2EDuration="7.199695392s" podCreationTimestamp="2025-10-03 13:15:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:15:56.025109728 +0000 UTC m=+1491.823581912" watchObservedRunningTime="2025-10-03 13:16:00.199695392 +0000 UTC m=+1495.998167596" Oct 03 13:16:00 crc kubenswrapper[4578]: I1003 13:16:00.226105 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jr79q" Oct 03 13:16:00 crc kubenswrapper[4578]: I1003 13:16:00.415070 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jr79q"] Oct 03 13:16:01 crc kubenswrapper[4578]: I1003 13:16:01.988154 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jr79q" podUID="dfc26ec2-4854-49fe-9c7b-5c548482c092" containerName="registry-server" containerID="cri-o://b195ec676ca478886e5fd427c958ce41b1614a8a51667378b0ef4d9f531c04e1" gracePeriod=2 Oct 03 13:16:02 crc kubenswrapper[4578]: I1003 13:16:02.266246 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 03 13:16:02 crc kubenswrapper[4578]: I1003 13:16:02.312801 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 03 13:16:02 crc kubenswrapper[4578]: I1003 13:16:02.444427 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jr79q" Oct 03 13:16:02 crc kubenswrapper[4578]: I1003 13:16:02.547619 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfc26ec2-4854-49fe-9c7b-5c548482c092-utilities\") pod \"dfc26ec2-4854-49fe-9c7b-5c548482c092\" (UID: \"dfc26ec2-4854-49fe-9c7b-5c548482c092\") " Oct 03 13:16:02 crc kubenswrapper[4578]: I1003 13:16:02.547707 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfc26ec2-4854-49fe-9c7b-5c548482c092-catalog-content\") pod \"dfc26ec2-4854-49fe-9c7b-5c548482c092\" (UID: \"dfc26ec2-4854-49fe-9c7b-5c548482c092\") " Oct 03 13:16:02 crc kubenswrapper[4578]: I1003 13:16:02.547771 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mqrb\" (UniqueName: \"kubernetes.io/projected/dfc26ec2-4854-49fe-9c7b-5c548482c092-kube-api-access-5mqrb\") pod \"dfc26ec2-4854-49fe-9c7b-5c548482c092\" (UID: \"dfc26ec2-4854-49fe-9c7b-5c548482c092\") " Oct 03 13:16:02 crc kubenswrapper[4578]: I1003 13:16:02.548429 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfc26ec2-4854-49fe-9c7b-5c548482c092-utilities" (OuterVolumeSpecName: "utilities") pod "dfc26ec2-4854-49fe-9c7b-5c548482c092" (UID: "dfc26ec2-4854-49fe-9c7b-5c548482c092"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:16:02 crc kubenswrapper[4578]: I1003 13:16:02.553873 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfc26ec2-4854-49fe-9c7b-5c548482c092-kube-api-access-5mqrb" (OuterVolumeSpecName: "kube-api-access-5mqrb") pod "dfc26ec2-4854-49fe-9c7b-5c548482c092" (UID: "dfc26ec2-4854-49fe-9c7b-5c548482c092"). InnerVolumeSpecName "kube-api-access-5mqrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:16:02 crc kubenswrapper[4578]: I1003 13:16:02.593775 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfc26ec2-4854-49fe-9c7b-5c548482c092-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dfc26ec2-4854-49fe-9c7b-5c548482c092" (UID: "dfc26ec2-4854-49fe-9c7b-5c548482c092"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:16:02 crc kubenswrapper[4578]: I1003 13:16:02.650236 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfc26ec2-4854-49fe-9c7b-5c548482c092-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:02 crc kubenswrapper[4578]: I1003 13:16:02.650273 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfc26ec2-4854-49fe-9c7b-5c548482c092-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:02 crc kubenswrapper[4578]: I1003 13:16:02.650285 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mqrb\" (UniqueName: \"kubernetes.io/projected/dfc26ec2-4854-49fe-9c7b-5c548482c092-kube-api-access-5mqrb\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:02 crc kubenswrapper[4578]: I1003 13:16:02.999238 4578 generic.go:334] "Generic (PLEG): container finished" podID="dfc26ec2-4854-49fe-9c7b-5c548482c092" containerID="b195ec676ca478886e5fd427c958ce41b1614a8a51667378b0ef4d9f531c04e1" exitCode=0 Oct 03 13:16:03 crc kubenswrapper[4578]: I1003 13:16:02.999993 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr79q" event={"ID":"dfc26ec2-4854-49fe-9c7b-5c548482c092","Type":"ContainerDied","Data":"b195ec676ca478886e5fd427c958ce41b1614a8a51667378b0ef4d9f531c04e1"} Oct 03 13:16:03 crc kubenswrapper[4578]: I1003 13:16:03.000032 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr79q" event={"ID":"dfc26ec2-4854-49fe-9c7b-5c548482c092","Type":"ContainerDied","Data":"ab6192501db15b837d443d0197beb5c8ff34606042fc1de2c1f76955208ddace"} Oct 03 13:16:03 crc kubenswrapper[4578]: I1003 13:16:03.000054 4578 scope.go:117] "RemoveContainer" containerID="b195ec676ca478886e5fd427c958ce41b1614a8a51667378b0ef4d9f531c04e1" Oct 03 13:16:03 crc kubenswrapper[4578]: I1003 13:16:03.000823 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jr79q" Oct 03 13:16:03 crc kubenswrapper[4578]: I1003 13:16:03.030842 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jr79q"] Oct 03 13:16:03 crc kubenswrapper[4578]: I1003 13:16:03.035341 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 03 13:16:03 crc kubenswrapper[4578]: I1003 13:16:03.051277 4578 scope.go:117] "RemoveContainer" containerID="5b6f0c812e1d7bb01b5d21e551f2d5c1e6f6aa4f6d52951afd385933aa21247a" Oct 03 13:16:03 crc kubenswrapper[4578]: I1003 13:16:03.052485 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jr79q"] Oct 03 13:16:03 crc kubenswrapper[4578]: I1003 13:16:03.110922 4578 scope.go:117] "RemoveContainer" containerID="5545d3d880bbea8435e1ce43aad7f0a2d16f186af17ecc2bd65dc5fc2a672d26" Oct 03 13:16:03 crc kubenswrapper[4578]: I1003 13:16:03.134903 4578 scope.go:117] "RemoveContainer" containerID="b195ec676ca478886e5fd427c958ce41b1614a8a51667378b0ef4d9f531c04e1" Oct 03 13:16:03 crc kubenswrapper[4578]: E1003 13:16:03.135679 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b195ec676ca478886e5fd427c958ce41b1614a8a51667378b0ef4d9f531c04e1\": container with ID starting with b195ec676ca478886e5fd427c958ce41b1614a8a51667378b0ef4d9f531c04e1 not found: ID does not exist" containerID="b195ec676ca478886e5fd427c958ce41b1614a8a51667378b0ef4d9f531c04e1" Oct 03 13:16:03 crc kubenswrapper[4578]: I1003 13:16:03.135761 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b195ec676ca478886e5fd427c958ce41b1614a8a51667378b0ef4d9f531c04e1"} err="failed to get container status \"b195ec676ca478886e5fd427c958ce41b1614a8a51667378b0ef4d9f531c04e1\": rpc error: code = NotFound desc = could not find container \"b195ec676ca478886e5fd427c958ce41b1614a8a51667378b0ef4d9f531c04e1\": container with ID starting with b195ec676ca478886e5fd427c958ce41b1614a8a51667378b0ef4d9f531c04e1 not found: ID does not exist" Oct 03 13:16:03 crc kubenswrapper[4578]: I1003 13:16:03.135822 4578 scope.go:117] "RemoveContainer" containerID="5b6f0c812e1d7bb01b5d21e551f2d5c1e6f6aa4f6d52951afd385933aa21247a" Oct 03 13:16:03 crc kubenswrapper[4578]: E1003 13:16:03.136474 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b6f0c812e1d7bb01b5d21e551f2d5c1e6f6aa4f6d52951afd385933aa21247a\": container with ID starting with 5b6f0c812e1d7bb01b5d21e551f2d5c1e6f6aa4f6d52951afd385933aa21247a not found: ID does not exist" containerID="5b6f0c812e1d7bb01b5d21e551f2d5c1e6f6aa4f6d52951afd385933aa21247a" Oct 03 13:16:03 crc kubenswrapper[4578]: I1003 13:16:03.136522 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b6f0c812e1d7bb01b5d21e551f2d5c1e6f6aa4f6d52951afd385933aa21247a"} err="failed to get container status \"5b6f0c812e1d7bb01b5d21e551f2d5c1e6f6aa4f6d52951afd385933aa21247a\": rpc error: code = NotFound desc = could not find container \"5b6f0c812e1d7bb01b5d21e551f2d5c1e6f6aa4f6d52951afd385933aa21247a\": container with ID starting with 5b6f0c812e1d7bb01b5d21e551f2d5c1e6f6aa4f6d52951afd385933aa21247a not found: ID does not exist" Oct 03 13:16:03 crc kubenswrapper[4578]: I1003 13:16:03.136550 4578 scope.go:117] "RemoveContainer" containerID="5545d3d880bbea8435e1ce43aad7f0a2d16f186af17ecc2bd65dc5fc2a672d26" Oct 03 13:16:03 crc kubenswrapper[4578]: E1003 13:16:03.136938 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5545d3d880bbea8435e1ce43aad7f0a2d16f186af17ecc2bd65dc5fc2a672d26\": container with ID starting with 5545d3d880bbea8435e1ce43aad7f0a2d16f186af17ecc2bd65dc5fc2a672d26 not found: ID does not exist" containerID="5545d3d880bbea8435e1ce43aad7f0a2d16f186af17ecc2bd65dc5fc2a672d26" Oct 03 13:16:03 crc kubenswrapper[4578]: I1003 13:16:03.136958 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5545d3d880bbea8435e1ce43aad7f0a2d16f186af17ecc2bd65dc5fc2a672d26"} err="failed to get container status \"5545d3d880bbea8435e1ce43aad7f0a2d16f186af17ecc2bd65dc5fc2a672d26\": rpc error: code = NotFound desc = could not find container \"5545d3d880bbea8435e1ce43aad7f0a2d16f186af17ecc2bd65dc5fc2a672d26\": container with ID starting with 5545d3d880bbea8435e1ce43aad7f0a2d16f186af17ecc2bd65dc5fc2a672d26 not found: ID does not exist" Oct 03 13:16:04 crc kubenswrapper[4578]: I1003 13:16:04.429363 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 13:16:04 crc kubenswrapper[4578]: I1003 13:16:04.429419 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 03 13:16:04 crc kubenswrapper[4578]: I1003 13:16:04.440005 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 03 13:16:04 crc kubenswrapper[4578]: I1003 13:16:04.440835 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 03 13:16:04 crc kubenswrapper[4578]: I1003 13:16:04.927984 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfc26ec2-4854-49fe-9c7b-5c548482c092" path="/var/lib/kubelet/pods/dfc26ec2-4854-49fe-9c7b-5c548482c092/volumes" Oct 03 13:16:05 crc kubenswrapper[4578]: I1003 13:16:05.093037 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:16:05 crc kubenswrapper[4578]: I1003 13:16:05.093295 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:16:05 crc kubenswrapper[4578]: I1003 13:16:05.093458 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 13:16:05 crc kubenswrapper[4578]: I1003 13:16:05.094362 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95"} pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:16:05 crc kubenswrapper[4578]: I1003 13:16:05.094527 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" containerID="cri-o://e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" gracePeriod=600 Oct 03 13:16:05 crc kubenswrapper[4578]: E1003 13:16:05.217729 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:16:05 crc kubenswrapper[4578]: I1003 13:16:05.439849 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="25b861ed-9dfe-4494-96a0-03fa994f1775" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 13:16:05 crc kubenswrapper[4578]: I1003 13:16:05.441169 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="25b861ed-9dfe-4494-96a0-03fa994f1775" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:16:05 crc kubenswrapper[4578]: I1003 13:16:05.450823 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="89f4ced5-a0c4-4b41-9c87-64377d0fc4bd" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 03 13:16:05 crc kubenswrapper[4578]: I1003 13:16:05.450829 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="89f4ced5-a0c4-4b41-9c87-64377d0fc4bd" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 03 13:16:05 crc kubenswrapper[4578]: I1003 13:16:05.940268 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 03 13:16:06 crc kubenswrapper[4578]: I1003 13:16:06.043064 4578 generic.go:334] "Generic (PLEG): container finished" podID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" exitCode=0 Oct 03 13:16:06 crc kubenswrapper[4578]: I1003 13:16:06.043112 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerDied","Data":"e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95"} Oct 03 13:16:06 crc kubenswrapper[4578]: I1003 13:16:06.043150 4578 scope.go:117] "RemoveContainer" containerID="ad105f3cc883a1c5d61b8c9107a9b7139beb580cccd8fece952ff237bc9b856e" Oct 03 13:16:06 crc kubenswrapper[4578]: I1003 13:16:06.043855 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:16:06 crc kubenswrapper[4578]: E1003 13:16:06.044123 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:16:14 crc kubenswrapper[4578]: I1003 13:16:14.434290 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 03 13:16:14 crc kubenswrapper[4578]: I1003 13:16:14.435839 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 03 13:16:14 crc kubenswrapper[4578]: I1003 13:16:14.435966 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 03 13:16:14 crc kubenswrapper[4578]: I1003 13:16:14.456106 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 03 13:16:14 crc kubenswrapper[4578]: I1003 13:16:14.458462 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 03 13:16:14 crc kubenswrapper[4578]: I1003 13:16:14.519001 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 03 13:16:14 crc kubenswrapper[4578]: I1003 13:16:14.661850 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 03 13:16:15 crc kubenswrapper[4578]: I1003 13:16:15.123330 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 03 13:16:15 crc kubenswrapper[4578]: I1003 13:16:15.131691 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 03 13:16:15 crc kubenswrapper[4578]: I1003 13:16:15.133519 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 03 13:16:20 crc kubenswrapper[4578]: I1003 13:16:20.909581 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:16:20 crc kubenswrapper[4578]: E1003 13:16:20.910359 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:16:23 crc kubenswrapper[4578]: I1003 13:16:23.606068 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 13:16:25 crc kubenswrapper[4578]: I1003 13:16:25.181244 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 13:16:27 crc kubenswrapper[4578]: I1003 13:16:27.942650 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="c196bfb1-793e-4e04-9602-db06886385ad" containerName="rabbitmq" containerID="cri-o://41aac4bc65fb0d271edee3e21b8ce5e0add0172a06237301735746d616fb09e3" gracePeriod=604796 Oct 03 13:16:29 crc kubenswrapper[4578]: I1003 13:16:29.360966 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="79fcccfa-6df4-41f2-965b-357e5a7984ab" containerName="rabbitmq" containerID="cri-o://6c285b7d81188b6d06e8bd70cfd27aeb562c9da338b801699ea591e8e5889ee9" gracePeriod=604796 Oct 03 13:16:30 crc kubenswrapper[4578]: I1003 13:16:30.055179 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="c196bfb1-793e-4e04-9602-db06886385ad" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.97:5671: connect: connection refused" Oct 03 13:16:30 crc kubenswrapper[4578]: I1003 13:16:30.461273 4578 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="79fcccfa-6df4-41f2-965b-357e5a7984ab" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Oct 03 13:16:31 crc kubenswrapper[4578]: I1003 13:16:31.909349 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:16:31 crc kubenswrapper[4578]: E1003 13:16:31.909891 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.306150 4578 generic.go:334] "Generic (PLEG): container finished" podID="c196bfb1-793e-4e04-9602-db06886385ad" containerID="41aac4bc65fb0d271edee3e21b8ce5e0add0172a06237301735746d616fb09e3" exitCode=0 Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.306230 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c196bfb1-793e-4e04-9602-db06886385ad","Type":"ContainerDied","Data":"41aac4bc65fb0d271edee3e21b8ce5e0add0172a06237301735746d616fb09e3"} Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.538889 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.638770 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c196bfb1-793e-4e04-9602-db06886385ad-server-conf\") pod \"c196bfb1-793e-4e04-9602-db06886385ad\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.638844 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c196bfb1-793e-4e04-9602-db06886385ad-pod-info\") pod \"c196bfb1-793e-4e04-9602-db06886385ad\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.638877 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"c196bfb1-793e-4e04-9602-db06886385ad\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.638915 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-confd\") pod \"c196bfb1-793e-4e04-9602-db06886385ad\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.638941 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c196bfb1-793e-4e04-9602-db06886385ad-config-data\") pod \"c196bfb1-793e-4e04-9602-db06886385ad\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.638970 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bh8vr\" (UniqueName: \"kubernetes.io/projected/c196bfb1-793e-4e04-9602-db06886385ad-kube-api-access-bh8vr\") pod \"c196bfb1-793e-4e04-9602-db06886385ad\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.638987 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c196bfb1-793e-4e04-9602-db06886385ad-plugins-conf\") pod \"c196bfb1-793e-4e04-9602-db06886385ad\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.639006 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-tls\") pod \"c196bfb1-793e-4e04-9602-db06886385ad\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.639036 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-erlang-cookie\") pod \"c196bfb1-793e-4e04-9602-db06886385ad\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.639085 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-plugins\") pod \"c196bfb1-793e-4e04-9602-db06886385ad\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.639160 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c196bfb1-793e-4e04-9602-db06886385ad-erlang-cookie-secret\") pod \"c196bfb1-793e-4e04-9602-db06886385ad\" (UID: \"c196bfb1-793e-4e04-9602-db06886385ad\") " Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.649852 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c196bfb1-793e-4e04-9602-db06886385ad-kube-api-access-bh8vr" (OuterVolumeSpecName: "kube-api-access-bh8vr") pod "c196bfb1-793e-4e04-9602-db06886385ad" (UID: "c196bfb1-793e-4e04-9602-db06886385ad"). InnerVolumeSpecName "kube-api-access-bh8vr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.667655 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c196bfb1-793e-4e04-9602-db06886385ad-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "c196bfb1-793e-4e04-9602-db06886385ad" (UID: "c196bfb1-793e-4e04-9602-db06886385ad"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.668139 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "c196bfb1-793e-4e04-9602-db06886385ad" (UID: "c196bfb1-793e-4e04-9602-db06886385ad"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.672078 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "c196bfb1-793e-4e04-9602-db06886385ad" (UID: "c196bfb1-793e-4e04-9602-db06886385ad"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.674148 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c196bfb1-793e-4e04-9602-db06886385ad-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "c196bfb1-793e-4e04-9602-db06886385ad" (UID: "c196bfb1-793e-4e04-9602-db06886385ad"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.679294 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "c196bfb1-793e-4e04-9602-db06886385ad" (UID: "c196bfb1-793e-4e04-9602-db06886385ad"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.682884 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "c196bfb1-793e-4e04-9602-db06886385ad" (UID: "c196bfb1-793e-4e04-9602-db06886385ad"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.687956 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c196bfb1-793e-4e04-9602-db06886385ad-pod-info" (OuterVolumeSpecName: "pod-info") pod "c196bfb1-793e-4e04-9602-db06886385ad" (UID: "c196bfb1-793e-4e04-9602-db06886385ad"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.695178 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c196bfb1-793e-4e04-9602-db06886385ad-config-data" (OuterVolumeSpecName: "config-data") pod "c196bfb1-793e-4e04-9602-db06886385ad" (UID: "c196bfb1-793e-4e04-9602-db06886385ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.723832 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c196bfb1-793e-4e04-9602-db06886385ad-server-conf" (OuterVolumeSpecName: "server-conf") pod "c196bfb1-793e-4e04-9602-db06886385ad" (UID: "c196bfb1-793e-4e04-9602-db06886385ad"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.740768 4578 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c196bfb1-793e-4e04-9602-db06886385ad-server-conf\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.740797 4578 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c196bfb1-793e-4e04-9602-db06886385ad-pod-info\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.740824 4578 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.740833 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c196bfb1-793e-4e04-9602-db06886385ad-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.740844 4578 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c196bfb1-793e-4e04-9602-db06886385ad-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.740854 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bh8vr\" (UniqueName: \"kubernetes.io/projected/c196bfb1-793e-4e04-9602-db06886385ad-kube-api-access-bh8vr\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.740863 4578 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.740873 4578 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.740883 4578 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.740892 4578 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c196bfb1-793e-4e04-9602-db06886385ad-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.842848 4578 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.843140 4578 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.861348 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "c196bfb1-793e-4e04-9602-db06886385ad" (UID: "c196bfb1-793e-4e04-9602-db06886385ad"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:16:34 crc kubenswrapper[4578]: I1003 13:16:34.944856 4578 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c196bfb1-793e-4e04-9602-db06886385ad-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.316732 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c196bfb1-793e-4e04-9602-db06886385ad","Type":"ContainerDied","Data":"b26db1e2f552b04b5d2c308687c6df6b76bbdcb0e3ef58559e8711050d5b6547"} Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.316782 4578 scope.go:117] "RemoveContainer" containerID="41aac4bc65fb0d271edee3e21b8ce5e0add0172a06237301735746d616fb09e3" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.316911 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.339476 4578 scope.go:117] "RemoveContainer" containerID="d3c16cfe0b8487b875181d1c37dc0cd7130f888f36465be2bb67dafa2f43ecaa" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.351437 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.363451 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.414313 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 13:16:35 crc kubenswrapper[4578]: E1003 13:16:35.414716 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfc26ec2-4854-49fe-9c7b-5c548482c092" containerName="extract-content" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.414742 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfc26ec2-4854-49fe-9c7b-5c548482c092" containerName="extract-content" Oct 03 13:16:35 crc kubenswrapper[4578]: E1003 13:16:35.414755 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfc26ec2-4854-49fe-9c7b-5c548482c092" containerName="registry-server" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.414761 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfc26ec2-4854-49fe-9c7b-5c548482c092" containerName="registry-server" Oct 03 13:16:35 crc kubenswrapper[4578]: E1003 13:16:35.414778 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c196bfb1-793e-4e04-9602-db06886385ad" containerName="setup-container" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.414784 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="c196bfb1-793e-4e04-9602-db06886385ad" containerName="setup-container" Oct 03 13:16:35 crc kubenswrapper[4578]: E1003 13:16:35.414797 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfc26ec2-4854-49fe-9c7b-5c548482c092" containerName="extract-utilities" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.414803 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfc26ec2-4854-49fe-9c7b-5c548482c092" containerName="extract-utilities" Oct 03 13:16:35 crc kubenswrapper[4578]: E1003 13:16:35.414812 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c196bfb1-793e-4e04-9602-db06886385ad" containerName="rabbitmq" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.414819 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="c196bfb1-793e-4e04-9602-db06886385ad" containerName="rabbitmq" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.415014 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="c196bfb1-793e-4e04-9602-db06886385ad" containerName="rabbitmq" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.415031 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfc26ec2-4854-49fe-9c7b-5c548482c092" containerName="registry-server" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.416089 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.442558 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.442826 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.442960 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.446152 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-nwbq9" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.446428 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.446589 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.446842 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.486053 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.556046 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dc895ffd-a4ba-497b-8600-c8491f007547-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.556120 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dc895ffd-a4ba-497b-8600-c8491f007547-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.556142 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc895ffd-a4ba-497b-8600-c8491f007547-config-data\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.556157 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dc895ffd-a4ba-497b-8600-c8491f007547-pod-info\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.556176 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dc895ffd-a4ba-497b-8600-c8491f007547-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.556220 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fsk6\" (UniqueName: \"kubernetes.io/projected/dc895ffd-a4ba-497b-8600-c8491f007547-kube-api-access-8fsk6\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.556255 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.556282 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dc895ffd-a4ba-497b-8600-c8491f007547-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.556316 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dc895ffd-a4ba-497b-8600-c8491f007547-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.556349 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dc895ffd-a4ba-497b-8600-c8491f007547-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.556384 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dc895ffd-a4ba-497b-8600-c8491f007547-server-conf\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.657748 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dc895ffd-a4ba-497b-8600-c8491f007547-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.658366 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dc895ffd-a4ba-497b-8600-c8491f007547-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.659202 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dc895ffd-a4ba-497b-8600-c8491f007547-server-conf\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.659251 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dc895ffd-a4ba-497b-8600-c8491f007547-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.659333 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dc895ffd-a4ba-497b-8600-c8491f007547-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.659360 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc895ffd-a4ba-497b-8600-c8491f007547-config-data\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.659384 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dc895ffd-a4ba-497b-8600-c8491f007547-pod-info\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.659414 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dc895ffd-a4ba-497b-8600-c8491f007547-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.659488 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fsk6\" (UniqueName: \"kubernetes.io/projected/dc895ffd-a4ba-497b-8600-c8491f007547-kube-api-access-8fsk6\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.659548 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.659586 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dc895ffd-a4ba-497b-8600-c8491f007547-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.659120 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dc895ffd-a4ba-497b-8600-c8491f007547-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.658307 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dc895ffd-a4ba-497b-8600-c8491f007547-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.661139 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dc895ffd-a4ba-497b-8600-c8491f007547-server-conf\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.661824 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dc895ffd-a4ba-497b-8600-c8491f007547-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.661925 4578 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.662106 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc895ffd-a4ba-497b-8600-c8491f007547-config-data\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.666172 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dc895ffd-a4ba-497b-8600-c8491f007547-pod-info\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.668956 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dc895ffd-a4ba-497b-8600-c8491f007547-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.673139 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dc895ffd-a4ba-497b-8600-c8491f007547-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.699990 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dc895ffd-a4ba-497b-8600-c8491f007547-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.703451 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fsk6\" (UniqueName: \"kubernetes.io/projected/dc895ffd-a4ba-497b-8600-c8491f007547-kube-api-access-8fsk6\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:35 crc kubenswrapper[4578]: I1003 13:16:35.803795 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"dc895ffd-a4ba-497b-8600-c8491f007547\") " pod="openstack/rabbitmq-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.005025 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.036973 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.067843 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79fcccfa-6df4-41f2-965b-357e5a7984ab-config-data\") pod \"79fcccfa-6df4-41f2-965b-357e5a7984ab\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.067885 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"79fcccfa-6df4-41f2-965b-357e5a7984ab\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.067906 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/79fcccfa-6df4-41f2-965b-357e5a7984ab-plugins-conf\") pod \"79fcccfa-6df4-41f2-965b-357e5a7984ab\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.067932 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-plugins\") pod \"79fcccfa-6df4-41f2-965b-357e5a7984ab\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.067984 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-tls\") pod \"79fcccfa-6df4-41f2-965b-357e5a7984ab\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.068014 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-confd\") pod \"79fcccfa-6df4-41f2-965b-357e5a7984ab\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.068037 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/79fcccfa-6df4-41f2-965b-357e5a7984ab-pod-info\") pod \"79fcccfa-6df4-41f2-965b-357e5a7984ab\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.068054 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/79fcccfa-6df4-41f2-965b-357e5a7984ab-server-conf\") pod \"79fcccfa-6df4-41f2-965b-357e5a7984ab\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.068109 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9zzn\" (UniqueName: \"kubernetes.io/projected/79fcccfa-6df4-41f2-965b-357e5a7984ab-kube-api-access-z9zzn\") pod \"79fcccfa-6df4-41f2-965b-357e5a7984ab\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.068174 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/79fcccfa-6df4-41f2-965b-357e5a7984ab-erlang-cookie-secret\") pod \"79fcccfa-6df4-41f2-965b-357e5a7984ab\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.068245 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-erlang-cookie\") pod \"79fcccfa-6df4-41f2-965b-357e5a7984ab\" (UID: \"79fcccfa-6df4-41f2-965b-357e5a7984ab\") " Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.080420 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79fcccfa-6df4-41f2-965b-357e5a7984ab-kube-api-access-z9zzn" (OuterVolumeSpecName: "kube-api-access-z9zzn") pod "79fcccfa-6df4-41f2-965b-357e5a7984ab" (UID: "79fcccfa-6df4-41f2-965b-357e5a7984ab"). InnerVolumeSpecName "kube-api-access-z9zzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.081092 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79fcccfa-6df4-41f2-965b-357e5a7984ab-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "79fcccfa-6df4-41f2-965b-357e5a7984ab" (UID: "79fcccfa-6df4-41f2-965b-357e5a7984ab"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.083736 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "79fcccfa-6df4-41f2-965b-357e5a7984ab" (UID: "79fcccfa-6df4-41f2-965b-357e5a7984ab"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.090419 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "79fcccfa-6df4-41f2-965b-357e5a7984ab" (UID: "79fcccfa-6df4-41f2-965b-357e5a7984ab"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.091566 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "79fcccfa-6df4-41f2-965b-357e5a7984ab" (UID: "79fcccfa-6df4-41f2-965b-357e5a7984ab"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.094444 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "79fcccfa-6df4-41f2-965b-357e5a7984ab" (UID: "79fcccfa-6df4-41f2-965b-357e5a7984ab"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.099379 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79fcccfa-6df4-41f2-965b-357e5a7984ab-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "79fcccfa-6df4-41f2-965b-357e5a7984ab" (UID: "79fcccfa-6df4-41f2-965b-357e5a7984ab"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.103901 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/79fcccfa-6df4-41f2-965b-357e5a7984ab-pod-info" (OuterVolumeSpecName: "pod-info") pod "79fcccfa-6df4-41f2-965b-357e5a7984ab" (UID: "79fcccfa-6df4-41f2-965b-357e5a7984ab"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.129334 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79fcccfa-6df4-41f2-965b-357e5a7984ab-config-data" (OuterVolumeSpecName: "config-data") pod "79fcccfa-6df4-41f2-965b-357e5a7984ab" (UID: "79fcccfa-6df4-41f2-965b-357e5a7984ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.170143 4578 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.170175 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/79fcccfa-6df4-41f2-965b-357e5a7984ab-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.170199 4578 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.170211 4578 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/79fcccfa-6df4-41f2-965b-357e5a7984ab-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.170223 4578 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.170234 4578 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.170245 4578 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/79fcccfa-6df4-41f2-965b-357e5a7984ab-pod-info\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.170254 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9zzn\" (UniqueName: \"kubernetes.io/projected/79fcccfa-6df4-41f2-965b-357e5a7984ab-kube-api-access-z9zzn\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.170262 4578 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/79fcccfa-6df4-41f2-965b-357e5a7984ab-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.213615 4578 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.248486 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79fcccfa-6df4-41f2-965b-357e5a7984ab-server-conf" (OuterVolumeSpecName: "server-conf") pod "79fcccfa-6df4-41f2-965b-357e5a7984ab" (UID: "79fcccfa-6df4-41f2-965b-357e5a7984ab"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.273572 4578 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.273897 4578 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/79fcccfa-6df4-41f2-965b-357e5a7984ab-server-conf\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.302111 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "79fcccfa-6df4-41f2-965b-357e5a7984ab" (UID: "79fcccfa-6df4-41f2-965b-357e5a7984ab"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.375440 4578 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/79fcccfa-6df4-41f2-965b-357e5a7984ab-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.401935 4578 generic.go:334] "Generic (PLEG): container finished" podID="79fcccfa-6df4-41f2-965b-357e5a7984ab" containerID="6c285b7d81188b6d06e8bd70cfd27aeb562c9da338b801699ea591e8e5889ee9" exitCode=0 Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.402047 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.402045 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"79fcccfa-6df4-41f2-965b-357e5a7984ab","Type":"ContainerDied","Data":"6c285b7d81188b6d06e8bd70cfd27aeb562c9da338b801699ea591e8e5889ee9"} Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.403080 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"79fcccfa-6df4-41f2-965b-357e5a7984ab","Type":"ContainerDied","Data":"1c17bc81fb7e6a865cc8a53c6027ca55b08892125ecd4876ede2e6717a1c23c3"} Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.403116 4578 scope.go:117] "RemoveContainer" containerID="6c285b7d81188b6d06e8bd70cfd27aeb562c9da338b801699ea591e8e5889ee9" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.456665 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.463204 4578 scope.go:117] "RemoveContainer" containerID="ff3fb681d8b8aa234547e28273c535c33bdeabb40fca2a2d071b337660eef7cd" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.464369 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.491431 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 13:16:36 crc kubenswrapper[4578]: E1003 13:16:36.492081 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79fcccfa-6df4-41f2-965b-357e5a7984ab" containerName="rabbitmq" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.492099 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="79fcccfa-6df4-41f2-965b-357e5a7984ab" containerName="rabbitmq" Oct 03 13:16:36 crc kubenswrapper[4578]: E1003 13:16:36.492122 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79fcccfa-6df4-41f2-965b-357e5a7984ab" containerName="setup-container" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.492129 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="79fcccfa-6df4-41f2-965b-357e5a7984ab" containerName="setup-container" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.492370 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="79fcccfa-6df4-41f2-965b-357e5a7984ab" containerName="rabbitmq" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.522609 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.541029 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-wxlrl" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.541262 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.541423 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.541536 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.542005 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.564146 4578 scope.go:117] "RemoveContainer" containerID="6c285b7d81188b6d06e8bd70cfd27aeb562c9da338b801699ea591e8e5889ee9" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.564622 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.564714 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.564846 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 03 13:16:36 crc kubenswrapper[4578]: E1003 13:16:36.569842 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c285b7d81188b6d06e8bd70cfd27aeb562c9da338b801699ea591e8e5889ee9\": container with ID starting with 6c285b7d81188b6d06e8bd70cfd27aeb562c9da338b801699ea591e8e5889ee9 not found: ID does not exist" containerID="6c285b7d81188b6d06e8bd70cfd27aeb562c9da338b801699ea591e8e5889ee9" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.571659 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c285b7d81188b6d06e8bd70cfd27aeb562c9da338b801699ea591e8e5889ee9"} err="failed to get container status \"6c285b7d81188b6d06e8bd70cfd27aeb562c9da338b801699ea591e8e5889ee9\": rpc error: code = NotFound desc = could not find container \"6c285b7d81188b6d06e8bd70cfd27aeb562c9da338b801699ea591e8e5889ee9\": container with ID starting with 6c285b7d81188b6d06e8bd70cfd27aeb562c9da338b801699ea591e8e5889ee9 not found: ID does not exist" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.571764 4578 scope.go:117] "RemoveContainer" containerID="ff3fb681d8b8aa234547e28273c535c33bdeabb40fca2a2d071b337660eef7cd" Oct 03 13:16:36 crc kubenswrapper[4578]: E1003 13:16:36.580553 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff3fb681d8b8aa234547e28273c535c33bdeabb40fca2a2d071b337660eef7cd\": container with ID starting with ff3fb681d8b8aa234547e28273c535c33bdeabb40fca2a2d071b337660eef7cd not found: ID does not exist" containerID="ff3fb681d8b8aa234547e28273c535c33bdeabb40fca2a2d071b337660eef7cd" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.580736 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff3fb681d8b8aa234547e28273c535c33bdeabb40fca2a2d071b337660eef7cd"} err="failed to get container status \"ff3fb681d8b8aa234547e28273c535c33bdeabb40fca2a2d071b337660eef7cd\": rpc error: code = NotFound desc = could not find container \"ff3fb681d8b8aa234547e28273c535c33bdeabb40fca2a2d071b337660eef7cd\": container with ID starting with ff3fb681d8b8aa234547e28273c535c33bdeabb40fca2a2d071b337660eef7cd not found: ID does not exist" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.682606 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.683228 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/56e6269a-6d4b-4e78-9fef-7224277e227a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.683323 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.683350 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/56e6269a-6d4b-4e78-9fef-7224277e227a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.683425 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56e6269a-6d4b-4e78-9fef-7224277e227a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.683456 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/56e6269a-6d4b-4e78-9fef-7224277e227a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.683479 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/56e6269a-6d4b-4e78-9fef-7224277e227a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.683501 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/56e6269a-6d4b-4e78-9fef-7224277e227a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.683517 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/56e6269a-6d4b-4e78-9fef-7224277e227a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.683532 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/56e6269a-6d4b-4e78-9fef-7224277e227a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.683550 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/56e6269a-6d4b-4e78-9fef-7224277e227a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.683581 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pq5p\" (UniqueName: \"kubernetes.io/projected/56e6269a-6d4b-4e78-9fef-7224277e227a-kube-api-access-2pq5p\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.785590 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/56e6269a-6d4b-4e78-9fef-7224277e227a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.785701 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.785724 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/56e6269a-6d4b-4e78-9fef-7224277e227a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.785796 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56e6269a-6d4b-4e78-9fef-7224277e227a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.785838 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/56e6269a-6d4b-4e78-9fef-7224277e227a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.785868 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/56e6269a-6d4b-4e78-9fef-7224277e227a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.785899 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/56e6269a-6d4b-4e78-9fef-7224277e227a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.785920 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/56e6269a-6d4b-4e78-9fef-7224277e227a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.785935 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/56e6269a-6d4b-4e78-9fef-7224277e227a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.785955 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/56e6269a-6d4b-4e78-9fef-7224277e227a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.785988 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pq5p\" (UniqueName: \"kubernetes.io/projected/56e6269a-6d4b-4e78-9fef-7224277e227a-kube-api-access-2pq5p\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.787338 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/56e6269a-6d4b-4e78-9fef-7224277e227a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.787594 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/56e6269a-6d4b-4e78-9fef-7224277e227a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.788146 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/56e6269a-6d4b-4e78-9fef-7224277e227a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.788492 4578 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.790026 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/56e6269a-6d4b-4e78-9fef-7224277e227a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.790110 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/56e6269a-6d4b-4e78-9fef-7224277e227a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.793317 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/56e6269a-6d4b-4e78-9fef-7224277e227a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.796160 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/56e6269a-6d4b-4e78-9fef-7224277e227a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.798012 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56e6269a-6d4b-4e78-9fef-7224277e227a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.802345 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pq5p\" (UniqueName: \"kubernetes.io/projected/56e6269a-6d4b-4e78-9fef-7224277e227a-kube-api-access-2pq5p\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.802414 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/56e6269a-6d4b-4e78-9fef-7224277e227a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.833080 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"56e6269a-6d4b-4e78-9fef-7224277e227a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.919317 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79fcccfa-6df4-41f2-965b-357e5a7984ab" path="/var/lib/kubelet/pods/79fcccfa-6df4-41f2-965b-357e5a7984ab/volumes" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.921564 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c196bfb1-793e-4e04-9602-db06886385ad" path="/var/lib/kubelet/pods/c196bfb1-793e-4e04-9602-db06886385ad/volumes" Oct 03 13:16:36 crc kubenswrapper[4578]: I1003 13:16:36.925613 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:16:37 crc kubenswrapper[4578]: W1003 13:16:37.376003 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56e6269a_6d4b_4e78_9fef_7224277e227a.slice/crio-c781c444799f61d7bb58e98577903df749e43e836831cea75d0a05f3e423de42 WatchSource:0}: Error finding container c781c444799f61d7bb58e98577903df749e43e836831cea75d0a05f3e423de42: Status 404 returned error can't find the container with id c781c444799f61d7bb58e98577903df749e43e836831cea75d0a05f3e423de42 Oct 03 13:16:37 crc kubenswrapper[4578]: I1003 13:16:37.377828 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 03 13:16:37 crc kubenswrapper[4578]: I1003 13:16:37.429815 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dc895ffd-a4ba-497b-8600-c8491f007547","Type":"ContainerStarted","Data":"d58daa78d020e0a4ec7a5d3a3453fc8df6368a16fe001ea07379851cddc363eb"} Oct 03 13:16:37 crc kubenswrapper[4578]: I1003 13:16:37.434574 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"56e6269a-6d4b-4e78-9fef-7224277e227a","Type":"ContainerStarted","Data":"c781c444799f61d7bb58e98577903df749e43e836831cea75d0a05f3e423de42"} Oct 03 13:16:38 crc kubenswrapper[4578]: I1003 13:16:38.447205 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dc895ffd-a4ba-497b-8600-c8491f007547","Type":"ContainerStarted","Data":"ecab4e5f85830ff28505545233c66ae26d28339144e55c5c3c7711701de9625a"} Oct 03 13:16:39 crc kubenswrapper[4578]: I1003 13:16:39.455693 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"56e6269a-6d4b-4e78-9fef-7224277e227a","Type":"ContainerStarted","Data":"dd9b05e82ff535c48524a993e18b6107ba2e91bd1b07ba3233ef5225629ba8a6"} Oct 03 13:16:40 crc kubenswrapper[4578]: I1003 13:16:40.841039 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-5sqfz"] Oct 03 13:16:40 crc kubenswrapper[4578]: I1003 13:16:40.848107 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:40 crc kubenswrapper[4578]: I1003 13:16:40.856151 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 03 13:16:40 crc kubenswrapper[4578]: I1003 13:16:40.861498 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-5sqfz"] Oct 03 13:16:40 crc kubenswrapper[4578]: I1003 13:16:40.964105 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:40 crc kubenswrapper[4578]: I1003 13:16:40.964457 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-config\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:40 crc kubenswrapper[4578]: I1003 13:16:40.964479 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-dns-svc\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:40 crc kubenswrapper[4578]: I1003 13:16:40.964504 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7kpt\" (UniqueName: \"kubernetes.io/projected/8d7937f8-46db-4b07-b64e-7cd7127c22eb-kube-api-access-n7kpt\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:40 crc kubenswrapper[4578]: I1003 13:16:40.964715 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:40 crc kubenswrapper[4578]: I1003 13:16:40.964815 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:40 crc kubenswrapper[4578]: I1003 13:16:40.964864 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:41 crc kubenswrapper[4578]: I1003 13:16:41.067384 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:41 crc kubenswrapper[4578]: I1003 13:16:41.067670 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:41 crc kubenswrapper[4578]: I1003 13:16:41.067777 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:41 crc kubenswrapper[4578]: I1003 13:16:41.067924 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:41 crc kubenswrapper[4578]: I1003 13:16:41.068030 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-config\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:41 crc kubenswrapper[4578]: I1003 13:16:41.068134 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-dns-svc\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:41 crc kubenswrapper[4578]: I1003 13:16:41.068583 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7kpt\" (UniqueName: \"kubernetes.io/projected/8d7937f8-46db-4b07-b64e-7cd7127c22eb-kube-api-access-n7kpt\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:41 crc kubenswrapper[4578]: I1003 13:16:41.068703 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:41 crc kubenswrapper[4578]: I1003 13:16:41.068531 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:41 crc kubenswrapper[4578]: I1003 13:16:41.068737 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:41 crc kubenswrapper[4578]: I1003 13:16:41.068931 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-config\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:41 crc kubenswrapper[4578]: I1003 13:16:41.069450 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-dns-svc\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:41 crc kubenswrapper[4578]: I1003 13:16:41.069466 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:41 crc kubenswrapper[4578]: I1003 13:16:41.094524 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7kpt\" (UniqueName: \"kubernetes.io/projected/8d7937f8-46db-4b07-b64e-7cd7127c22eb-kube-api-access-n7kpt\") pod \"dnsmasq-dns-67b789f86c-5sqfz\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:41 crc kubenswrapper[4578]: I1003 13:16:41.193639 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:41 crc kubenswrapper[4578]: I1003 13:16:41.661095 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-5sqfz"] Oct 03 13:16:41 crc kubenswrapper[4578]: W1003 13:16:41.666172 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d7937f8_46db_4b07_b64e_7cd7127c22eb.slice/crio-4c915eb07dc36e6fb6e4d43c4d915457c898ac19bb826d22a6467209578f84da WatchSource:0}: Error finding container 4c915eb07dc36e6fb6e4d43c4d915457c898ac19bb826d22a6467209578f84da: Status 404 returned error can't find the container with id 4c915eb07dc36e6fb6e4d43c4d915457c898ac19bb826d22a6467209578f84da Oct 03 13:16:42 crc kubenswrapper[4578]: I1003 13:16:42.494747 4578 generic.go:334] "Generic (PLEG): container finished" podID="8d7937f8-46db-4b07-b64e-7cd7127c22eb" containerID="09e27fdd4932910118d29a9321bff7912ad0e544e03a5bc526dc410e133e7049" exitCode=0 Oct 03 13:16:42 crc kubenswrapper[4578]: I1003 13:16:42.495203 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" event={"ID":"8d7937f8-46db-4b07-b64e-7cd7127c22eb","Type":"ContainerDied","Data":"09e27fdd4932910118d29a9321bff7912ad0e544e03a5bc526dc410e133e7049"} Oct 03 13:16:42 crc kubenswrapper[4578]: I1003 13:16:42.496099 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" event={"ID":"8d7937f8-46db-4b07-b64e-7cd7127c22eb","Type":"ContainerStarted","Data":"4c915eb07dc36e6fb6e4d43c4d915457c898ac19bb826d22a6467209578f84da"} Oct 03 13:16:43 crc kubenswrapper[4578]: I1003 13:16:43.506021 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" event={"ID":"8d7937f8-46db-4b07-b64e-7cd7127c22eb","Type":"ContainerStarted","Data":"f35464a42a83df88a6040ba9a2be005f94ab5689a935f37ebb45f9d11dee2bce"} Oct 03 13:16:43 crc kubenswrapper[4578]: I1003 13:16:43.506415 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:43 crc kubenswrapper[4578]: I1003 13:16:43.528830 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" podStartSLOduration=3.528803039 podStartE2EDuration="3.528803039s" podCreationTimestamp="2025-10-03 13:16:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:16:43.525548346 +0000 UTC m=+1539.324020540" watchObservedRunningTime="2025-10-03 13:16:43.528803039 +0000 UTC m=+1539.327275223" Oct 03 13:16:43 crc kubenswrapper[4578]: I1003 13:16:43.908955 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:16:43 crc kubenswrapper[4578]: E1003 13:16:43.909205 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.195816 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.306717 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-wrzpd"] Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.307003 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" podUID="6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df" containerName="dnsmasq-dns" containerID="cri-o://a930896ada2bd8eb0a5a065becdd4f556524d6303bd4f82d75e6654137c045d7" gracePeriod=10 Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.528150 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79dc84bdb7-82gmw"] Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.533212 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.541917 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79dc84bdb7-82gmw"] Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.580935 4578 generic.go:334] "Generic (PLEG): container finished" podID="6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df" containerID="a930896ada2bd8eb0a5a065becdd4f556524d6303bd4f82d75e6654137c045d7" exitCode=0 Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.580992 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" event={"ID":"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df","Type":"ContainerDied","Data":"a930896ada2bd8eb0a5a065becdd4f556524d6303bd4f82d75e6654137c045d7"} Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.680197 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-dns-svc\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.681013 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-config\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.681042 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-dns-swift-storage-0\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.681065 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-openstack-edpm-ipam\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.681114 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-ovsdbserver-nb\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.681142 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-ovsdbserver-sb\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.682992 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8cz7\" (UniqueName: \"kubernetes.io/projected/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-kube-api-access-g8cz7\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.784242 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-dns-svc\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.784372 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-config\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.784473 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-dns-swift-storage-0\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.784500 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-openstack-edpm-ipam\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.784544 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-ovsdbserver-nb\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.784570 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-ovsdbserver-sb\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.784595 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8cz7\" (UniqueName: \"kubernetes.io/projected/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-kube-api-access-g8cz7\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.786477 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-ovsdbserver-nb\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.786494 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-ovsdbserver-sb\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.786617 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-config\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.787048 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-openstack-edpm-ipam\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.787203 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-dns-svc\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.787352 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-dns-swift-storage-0\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.817437 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8cz7\" (UniqueName: \"kubernetes.io/projected/76bba854-23ba-4bf1-b4aa-db525ff8c5d8-kube-api-access-g8cz7\") pod \"dnsmasq-dns-79dc84bdb7-82gmw\" (UID: \"76bba854-23ba-4bf1-b4aa-db525ff8c5d8\") " pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.874934 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:51 crc kubenswrapper[4578]: I1003 13:16:51.969352 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.091184 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-config\") pod \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.091225 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-ovsdbserver-sb\") pod \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.091273 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-dns-svc\") pod \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.091337 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-dns-swift-storage-0\") pod \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.091376 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzrjc\" (UniqueName: \"kubernetes.io/projected/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-kube-api-access-jzrjc\") pod \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.091432 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-ovsdbserver-nb\") pod \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\" (UID: \"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df\") " Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.103953 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-kube-api-access-jzrjc" (OuterVolumeSpecName: "kube-api-access-jzrjc") pod "6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df" (UID: "6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df"). InnerVolumeSpecName "kube-api-access-jzrjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.175912 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df" (UID: "6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.187855 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-config" (OuterVolumeSpecName: "config") pod "6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df" (UID: "6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.197562 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.197607 4578 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.197624 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzrjc\" (UniqueName: \"kubernetes.io/projected/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-kube-api-access-jzrjc\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.198371 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79dc84bdb7-82gmw"] Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.202178 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df" (UID: "6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:16:52 crc kubenswrapper[4578]: W1003 13:16:52.202955 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76bba854_23ba_4bf1_b4aa_db525ff8c5d8.slice/crio-1431267ac4a80a657e59a7aaf90824ae0a0fed34d23675074c2dabe138fc5461 WatchSource:0}: Error finding container 1431267ac4a80a657e59a7aaf90824ae0a0fed34d23675074c2dabe138fc5461: Status 404 returned error can't find the container with id 1431267ac4a80a657e59a7aaf90824ae0a0fed34d23675074c2dabe138fc5461 Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.205233 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df" (UID: "6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.219444 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df" (UID: "6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.299694 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.299723 4578 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.299733 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.590909 4578 generic.go:334] "Generic (PLEG): container finished" podID="76bba854-23ba-4bf1-b4aa-db525ff8c5d8" containerID="4e9f0917b53c039a70dc7b89c545a6a244133638763681855aa304733ad8989e" exitCode=0 Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.590984 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" event={"ID":"76bba854-23ba-4bf1-b4aa-db525ff8c5d8","Type":"ContainerDied","Data":"4e9f0917b53c039a70dc7b89c545a6a244133638763681855aa304733ad8989e"} Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.591014 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" event={"ID":"76bba854-23ba-4bf1-b4aa-db525ff8c5d8","Type":"ContainerStarted","Data":"1431267ac4a80a657e59a7aaf90824ae0a0fed34d23675074c2dabe138fc5461"} Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.595614 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" event={"ID":"6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df","Type":"ContainerDied","Data":"8b545aac0e5015a402f4017420104302b2a9f29e984ad903ea3f04f22ddc2803"} Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.595693 4578 scope.go:117] "RemoveContainer" containerID="a930896ada2bd8eb0a5a065becdd4f556524d6303bd4f82d75e6654137c045d7" Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.595834 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-wrzpd" Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.798414 4578 scope.go:117] "RemoveContainer" containerID="6fa9e91c48b7d8258eabf25b71cd390230210d941e72e5a6c52307399d53284c" Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.832422 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-wrzpd"] Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.845133 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-wrzpd"] Oct 03 13:16:52 crc kubenswrapper[4578]: I1003 13:16:52.925822 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df" path="/var/lib/kubelet/pods/6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df/volumes" Oct 03 13:16:53 crc kubenswrapper[4578]: I1003 13:16:53.607084 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" event={"ID":"76bba854-23ba-4bf1-b4aa-db525ff8c5d8","Type":"ContainerStarted","Data":"de42b3170c56c2f31c2971a1fecd9e8086dc68f5e1b62d08dd18ea79402d4d19"} Oct 03 13:16:53 crc kubenswrapper[4578]: I1003 13:16:53.607321 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:16:53 crc kubenswrapper[4578]: I1003 13:16:53.628943 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" podStartSLOduration=2.628926823 podStartE2EDuration="2.628926823s" podCreationTimestamp="2025-10-03 13:16:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:16:53.625600898 +0000 UTC m=+1549.424073092" watchObservedRunningTime="2025-10-03 13:16:53.628926823 +0000 UTC m=+1549.427399007" Oct 03 13:16:55 crc kubenswrapper[4578]: I1003 13:16:55.908875 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:16:55 crc kubenswrapper[4578]: E1003 13:16:55.910479 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:17:01 crc kubenswrapper[4578]: I1003 13:17:01.877806 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79dc84bdb7-82gmw" Oct 03 13:17:01 crc kubenswrapper[4578]: I1003 13:17:01.951896 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-5sqfz"] Oct 03 13:17:01 crc kubenswrapper[4578]: I1003 13:17:01.952134 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" podUID="8d7937f8-46db-4b07-b64e-7cd7127c22eb" containerName="dnsmasq-dns" containerID="cri-o://f35464a42a83df88a6040ba9a2be005f94ab5689a935f37ebb45f9d11dee2bce" gracePeriod=10 Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.484494 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.588952 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7kpt\" (UniqueName: \"kubernetes.io/projected/8d7937f8-46db-4b07-b64e-7cd7127c22eb-kube-api-access-n7kpt\") pod \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.589057 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-ovsdbserver-nb\") pod \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.589109 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-ovsdbserver-sb\") pod \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.589170 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-dns-svc\") pod \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.589217 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-openstack-edpm-ipam\") pod \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.589245 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-dns-swift-storage-0\") pod \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.589358 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-config\") pod \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\" (UID: \"8d7937f8-46db-4b07-b64e-7cd7127c22eb\") " Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.612449 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d7937f8-46db-4b07-b64e-7cd7127c22eb-kube-api-access-n7kpt" (OuterVolumeSpecName: "kube-api-access-n7kpt") pod "8d7937f8-46db-4b07-b64e-7cd7127c22eb" (UID: "8d7937f8-46db-4b07-b64e-7cd7127c22eb"). InnerVolumeSpecName "kube-api-access-n7kpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.648770 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8d7937f8-46db-4b07-b64e-7cd7127c22eb" (UID: "8d7937f8-46db-4b07-b64e-7cd7127c22eb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.652109 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8d7937f8-46db-4b07-b64e-7cd7127c22eb" (UID: "8d7937f8-46db-4b07-b64e-7cd7127c22eb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.656361 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8d7937f8-46db-4b07-b64e-7cd7127c22eb" (UID: "8d7937f8-46db-4b07-b64e-7cd7127c22eb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.660087 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8d7937f8-46db-4b07-b64e-7cd7127c22eb" (UID: "8d7937f8-46db-4b07-b64e-7cd7127c22eb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.676383 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "8d7937f8-46db-4b07-b64e-7cd7127c22eb" (UID: "8d7937f8-46db-4b07-b64e-7cd7127c22eb"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.692138 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7kpt\" (UniqueName: \"kubernetes.io/projected/8d7937f8-46db-4b07-b64e-7cd7127c22eb-kube-api-access-n7kpt\") on node \"crc\" DevicePath \"\"" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.692181 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.692193 4578 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.692204 4578 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.692219 4578 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.692230 4578 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.696105 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-config" (OuterVolumeSpecName: "config") pod "8d7937f8-46db-4b07-b64e-7cd7127c22eb" (UID: "8d7937f8-46db-4b07-b64e-7cd7127c22eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.754872 4578 generic.go:334] "Generic (PLEG): container finished" podID="8d7937f8-46db-4b07-b64e-7cd7127c22eb" containerID="f35464a42a83df88a6040ba9a2be005f94ab5689a935f37ebb45f9d11dee2bce" exitCode=0 Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.754918 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" event={"ID":"8d7937f8-46db-4b07-b64e-7cd7127c22eb","Type":"ContainerDied","Data":"f35464a42a83df88a6040ba9a2be005f94ab5689a935f37ebb45f9d11dee2bce"} Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.754988 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" event={"ID":"8d7937f8-46db-4b07-b64e-7cd7127c22eb","Type":"ContainerDied","Data":"4c915eb07dc36e6fb6e4d43c4d915457c898ac19bb826d22a6467209578f84da"} Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.755007 4578 scope.go:117] "RemoveContainer" containerID="f35464a42a83df88a6040ba9a2be005f94ab5689a935f37ebb45f9d11dee2bce" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.755220 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-5sqfz" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.782755 4578 scope.go:117] "RemoveContainer" containerID="09e27fdd4932910118d29a9321bff7912ad0e544e03a5bc526dc410e133e7049" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.795063 4578 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d7937f8-46db-4b07-b64e-7cd7127c22eb-config\") on node \"crc\" DevicePath \"\"" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.816467 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-5sqfz"] Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.828268 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-5sqfz"] Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.835829 4578 scope.go:117] "RemoveContainer" containerID="f35464a42a83df88a6040ba9a2be005f94ab5689a935f37ebb45f9d11dee2bce" Oct 03 13:17:02 crc kubenswrapper[4578]: E1003 13:17:02.839800 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f35464a42a83df88a6040ba9a2be005f94ab5689a935f37ebb45f9d11dee2bce\": container with ID starting with f35464a42a83df88a6040ba9a2be005f94ab5689a935f37ebb45f9d11dee2bce not found: ID does not exist" containerID="f35464a42a83df88a6040ba9a2be005f94ab5689a935f37ebb45f9d11dee2bce" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.839854 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f35464a42a83df88a6040ba9a2be005f94ab5689a935f37ebb45f9d11dee2bce"} err="failed to get container status \"f35464a42a83df88a6040ba9a2be005f94ab5689a935f37ebb45f9d11dee2bce\": rpc error: code = NotFound desc = could not find container \"f35464a42a83df88a6040ba9a2be005f94ab5689a935f37ebb45f9d11dee2bce\": container with ID starting with f35464a42a83df88a6040ba9a2be005f94ab5689a935f37ebb45f9d11dee2bce not found: ID does not exist" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.839892 4578 scope.go:117] "RemoveContainer" containerID="09e27fdd4932910118d29a9321bff7912ad0e544e03a5bc526dc410e133e7049" Oct 03 13:17:02 crc kubenswrapper[4578]: E1003 13:17:02.843812 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09e27fdd4932910118d29a9321bff7912ad0e544e03a5bc526dc410e133e7049\": container with ID starting with 09e27fdd4932910118d29a9321bff7912ad0e544e03a5bc526dc410e133e7049 not found: ID does not exist" containerID="09e27fdd4932910118d29a9321bff7912ad0e544e03a5bc526dc410e133e7049" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.843861 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09e27fdd4932910118d29a9321bff7912ad0e544e03a5bc526dc410e133e7049"} err="failed to get container status \"09e27fdd4932910118d29a9321bff7912ad0e544e03a5bc526dc410e133e7049\": rpc error: code = NotFound desc = could not find container \"09e27fdd4932910118d29a9321bff7912ad0e544e03a5bc526dc410e133e7049\": container with ID starting with 09e27fdd4932910118d29a9321bff7912ad0e544e03a5bc526dc410e133e7049 not found: ID does not exist" Oct 03 13:17:02 crc kubenswrapper[4578]: I1003 13:17:02.921819 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d7937f8-46db-4b07-b64e-7cd7127c22eb" path="/var/lib/kubelet/pods/8d7937f8-46db-4b07-b64e-7cd7127c22eb/volumes" Oct 03 13:17:08 crc kubenswrapper[4578]: I1003 13:17:08.909368 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:17:08 crc kubenswrapper[4578]: E1003 13:17:08.910035 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:17:10 crc kubenswrapper[4578]: I1003 13:17:10.823997 4578 generic.go:334] "Generic (PLEG): container finished" podID="dc895ffd-a4ba-497b-8600-c8491f007547" containerID="ecab4e5f85830ff28505545233c66ae26d28339144e55c5c3c7711701de9625a" exitCode=0 Oct 03 13:17:10 crc kubenswrapper[4578]: I1003 13:17:10.824504 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dc895ffd-a4ba-497b-8600-c8491f007547","Type":"ContainerDied","Data":"ecab4e5f85830ff28505545233c66ae26d28339144e55c5c3c7711701de9625a"} Oct 03 13:17:10 crc kubenswrapper[4578]: I1003 13:17:10.827271 4578 generic.go:334] "Generic (PLEG): container finished" podID="56e6269a-6d4b-4e78-9fef-7224277e227a" containerID="dd9b05e82ff535c48524a993e18b6107ba2e91bd1b07ba3233ef5225629ba8a6" exitCode=0 Oct 03 13:17:10 crc kubenswrapper[4578]: I1003 13:17:10.827315 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"56e6269a-6d4b-4e78-9fef-7224277e227a","Type":"ContainerDied","Data":"dd9b05e82ff535c48524a993e18b6107ba2e91bd1b07ba3233ef5225629ba8a6"} Oct 03 13:17:11 crc kubenswrapper[4578]: I1003 13:17:11.841190 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"dc895ffd-a4ba-497b-8600-c8491f007547","Type":"ContainerStarted","Data":"3b1deef52a9ec96f736fdbf3d9010fb4a0796a3e2382390779ef065cbf2897df"} Oct 03 13:17:11 crc kubenswrapper[4578]: I1003 13:17:11.841703 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 03 13:17:11 crc kubenswrapper[4578]: I1003 13:17:11.842961 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"56e6269a-6d4b-4e78-9fef-7224277e227a","Type":"ContainerStarted","Data":"5a83d2b5ab91da9595472fb34c0ec488ad73323a70dd5c73a1fc282e77ec0d76"} Oct 03 13:17:11 crc kubenswrapper[4578]: I1003 13:17:11.843855 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:17:11 crc kubenswrapper[4578]: I1003 13:17:11.866652 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.866620167 podStartE2EDuration="36.866620167s" podCreationTimestamp="2025-10-03 13:16:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:17:11.861054803 +0000 UTC m=+1567.659526987" watchObservedRunningTime="2025-10-03 13:17:11.866620167 +0000 UTC m=+1567.665092351" Oct 03 13:17:11 crc kubenswrapper[4578]: I1003 13:17:11.892784 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=35.892766608 podStartE2EDuration="35.892766608s" podCreationTimestamp="2025-10-03 13:16:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:17:11.89090914 +0000 UTC m=+1567.689381324" watchObservedRunningTime="2025-10-03 13:17:11.892766608 +0000 UTC m=+1567.691238792" Oct 03 13:17:23 crc kubenswrapper[4578]: I1003 13:17:23.909769 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:17:23 crc kubenswrapper[4578]: E1003 13:17:23.910565 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.318828 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq"] Oct 03 13:17:25 crc kubenswrapper[4578]: E1003 13:17:25.319623 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d7937f8-46db-4b07-b64e-7cd7127c22eb" containerName="dnsmasq-dns" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.319658 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d7937f8-46db-4b07-b64e-7cd7127c22eb" containerName="dnsmasq-dns" Oct 03 13:17:25 crc kubenswrapper[4578]: E1003 13:17:25.319698 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df" containerName="init" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.319707 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df" containerName="init" Oct 03 13:17:25 crc kubenswrapper[4578]: E1003 13:17:25.319722 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d7937f8-46db-4b07-b64e-7cd7127c22eb" containerName="init" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.319730 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d7937f8-46db-4b07-b64e-7cd7127c22eb" containerName="init" Oct 03 13:17:25 crc kubenswrapper[4578]: E1003 13:17:25.319751 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df" containerName="dnsmasq-dns" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.319758 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df" containerName="dnsmasq-dns" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.319983 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fe75e3a-2d2b-4b6a-a7f4-521fe5f456df" containerName="dnsmasq-dns" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.320010 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d7937f8-46db-4b07-b64e-7cd7127c22eb" containerName="dnsmasq-dns" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.320771 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.326050 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.326119 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.326278 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-442mn" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.326424 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.347489 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq"] Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.521203 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgcqk\" (UniqueName: \"kubernetes.io/projected/6d166384-568f-42ea-a441-4d1df54fd5ce-kube-api-access-vgcqk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-457zq\" (UID: \"6d166384-568f-42ea-a441-4d1df54fd5ce\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.521272 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d166384-568f-42ea-a441-4d1df54fd5ce-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-457zq\" (UID: \"6d166384-568f-42ea-a441-4d1df54fd5ce\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.521346 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d166384-568f-42ea-a441-4d1df54fd5ce-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-457zq\" (UID: \"6d166384-568f-42ea-a441-4d1df54fd5ce\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.521413 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d166384-568f-42ea-a441-4d1df54fd5ce-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-457zq\" (UID: \"6d166384-568f-42ea-a441-4d1df54fd5ce\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.623648 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgcqk\" (UniqueName: \"kubernetes.io/projected/6d166384-568f-42ea-a441-4d1df54fd5ce-kube-api-access-vgcqk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-457zq\" (UID: \"6d166384-568f-42ea-a441-4d1df54fd5ce\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.623702 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d166384-568f-42ea-a441-4d1df54fd5ce-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-457zq\" (UID: \"6d166384-568f-42ea-a441-4d1df54fd5ce\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.623759 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d166384-568f-42ea-a441-4d1df54fd5ce-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-457zq\" (UID: \"6d166384-568f-42ea-a441-4d1df54fd5ce\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.623797 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d166384-568f-42ea-a441-4d1df54fd5ce-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-457zq\" (UID: \"6d166384-568f-42ea-a441-4d1df54fd5ce\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.629870 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d166384-568f-42ea-a441-4d1df54fd5ce-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-457zq\" (UID: \"6d166384-568f-42ea-a441-4d1df54fd5ce\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.633248 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d166384-568f-42ea-a441-4d1df54fd5ce-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-457zq\" (UID: \"6d166384-568f-42ea-a441-4d1df54fd5ce\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.638992 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d166384-568f-42ea-a441-4d1df54fd5ce-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-457zq\" (UID: \"6d166384-568f-42ea-a441-4d1df54fd5ce\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.640269 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgcqk\" (UniqueName: \"kubernetes.io/projected/6d166384-568f-42ea-a441-4d1df54fd5ce-kube-api-access-vgcqk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-457zq\" (UID: \"6d166384-568f-42ea-a441-4d1df54fd5ce\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" Oct 03 13:17:25 crc kubenswrapper[4578]: I1003 13:17:25.644170 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" Oct 03 13:17:26 crc kubenswrapper[4578]: I1003 13:17:26.041843 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 03 13:17:26 crc kubenswrapper[4578]: I1003 13:17:26.488303 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq"] Oct 03 13:17:26 crc kubenswrapper[4578]: W1003 13:17:26.495278 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d166384_568f_42ea_a441_4d1df54fd5ce.slice/crio-62b9a58aa21f584067d1c96d0a6b81becbab23a424a72e7e289318cba48d3d6e WatchSource:0}: Error finding container 62b9a58aa21f584067d1c96d0a6b81becbab23a424a72e7e289318cba48d3d6e: Status 404 returned error can't find the container with id 62b9a58aa21f584067d1c96d0a6b81becbab23a424a72e7e289318cba48d3d6e Oct 03 13:17:26 crc kubenswrapper[4578]: I1003 13:17:26.929255 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 03 13:17:27 crc kubenswrapper[4578]: I1003 13:17:27.017663 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" event={"ID":"6d166384-568f-42ea-a441-4d1df54fd5ce","Type":"ContainerStarted","Data":"62b9a58aa21f584067d1c96d0a6b81becbab23a424a72e7e289318cba48d3d6e"} Oct 03 13:17:34 crc kubenswrapper[4578]: I1003 13:17:34.920225 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:17:34 crc kubenswrapper[4578]: E1003 13:17:34.921025 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:17:41 crc kubenswrapper[4578]: I1003 13:17:41.159616 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" event={"ID":"6d166384-568f-42ea-a441-4d1df54fd5ce","Type":"ContainerStarted","Data":"c7ea4f5e7503367244dca159b9eacb78bdd9330bfe70bca45fa430ab30564834"} Oct 03 13:17:41 crc kubenswrapper[4578]: I1003 13:17:41.180593 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" podStartSLOduration=2.200573151 podStartE2EDuration="16.180572731s" podCreationTimestamp="2025-10-03 13:17:25 +0000 UTC" firstStartedPulling="2025-10-03 13:17:26.501325294 +0000 UTC m=+1582.299797478" lastFinishedPulling="2025-10-03 13:17:40.481324874 +0000 UTC m=+1596.279797058" observedRunningTime="2025-10-03 13:17:41.174840131 +0000 UTC m=+1596.973312335" watchObservedRunningTime="2025-10-03 13:17:41.180572731 +0000 UTC m=+1596.979044915" Oct 03 13:17:48 crc kubenswrapper[4578]: I1003 13:17:48.909352 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:17:48 crc kubenswrapper[4578]: E1003 13:17:48.910156 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:17:54 crc kubenswrapper[4578]: I1003 13:17:54.789618 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z22h2"] Oct 03 13:17:54 crc kubenswrapper[4578]: I1003 13:17:54.792252 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z22h2" Oct 03 13:17:54 crc kubenswrapper[4578]: I1003 13:17:54.805481 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z22h2"] Oct 03 13:17:54 crc kubenswrapper[4578]: I1003 13:17:54.831814 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62126af2-4cc0-4ccd-bb45-2520abe0f53d-utilities\") pod \"redhat-operators-z22h2\" (UID: \"62126af2-4cc0-4ccd-bb45-2520abe0f53d\") " pod="openshift-marketplace/redhat-operators-z22h2" Oct 03 13:17:54 crc kubenswrapper[4578]: I1003 13:17:54.832051 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xqp6\" (UniqueName: \"kubernetes.io/projected/62126af2-4cc0-4ccd-bb45-2520abe0f53d-kube-api-access-9xqp6\") pod \"redhat-operators-z22h2\" (UID: \"62126af2-4cc0-4ccd-bb45-2520abe0f53d\") " pod="openshift-marketplace/redhat-operators-z22h2" Oct 03 13:17:54 crc kubenswrapper[4578]: I1003 13:17:54.832174 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62126af2-4cc0-4ccd-bb45-2520abe0f53d-catalog-content\") pod \"redhat-operators-z22h2\" (UID: \"62126af2-4cc0-4ccd-bb45-2520abe0f53d\") " pod="openshift-marketplace/redhat-operators-z22h2" Oct 03 13:17:54 crc kubenswrapper[4578]: I1003 13:17:54.933915 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62126af2-4cc0-4ccd-bb45-2520abe0f53d-utilities\") pod \"redhat-operators-z22h2\" (UID: \"62126af2-4cc0-4ccd-bb45-2520abe0f53d\") " pod="openshift-marketplace/redhat-operators-z22h2" Oct 03 13:17:54 crc kubenswrapper[4578]: I1003 13:17:54.934061 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xqp6\" (UniqueName: \"kubernetes.io/projected/62126af2-4cc0-4ccd-bb45-2520abe0f53d-kube-api-access-9xqp6\") pod \"redhat-operators-z22h2\" (UID: \"62126af2-4cc0-4ccd-bb45-2520abe0f53d\") " pod="openshift-marketplace/redhat-operators-z22h2" Oct 03 13:17:54 crc kubenswrapper[4578]: I1003 13:17:54.934114 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62126af2-4cc0-4ccd-bb45-2520abe0f53d-catalog-content\") pod \"redhat-operators-z22h2\" (UID: \"62126af2-4cc0-4ccd-bb45-2520abe0f53d\") " pod="openshift-marketplace/redhat-operators-z22h2" Oct 03 13:17:54 crc kubenswrapper[4578]: I1003 13:17:54.934478 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62126af2-4cc0-4ccd-bb45-2520abe0f53d-utilities\") pod \"redhat-operators-z22h2\" (UID: \"62126af2-4cc0-4ccd-bb45-2520abe0f53d\") " pod="openshift-marketplace/redhat-operators-z22h2" Oct 03 13:17:54 crc kubenswrapper[4578]: I1003 13:17:54.934574 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62126af2-4cc0-4ccd-bb45-2520abe0f53d-catalog-content\") pod \"redhat-operators-z22h2\" (UID: \"62126af2-4cc0-4ccd-bb45-2520abe0f53d\") " pod="openshift-marketplace/redhat-operators-z22h2" Oct 03 13:17:54 crc kubenswrapper[4578]: I1003 13:17:54.955602 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xqp6\" (UniqueName: \"kubernetes.io/projected/62126af2-4cc0-4ccd-bb45-2520abe0f53d-kube-api-access-9xqp6\") pod \"redhat-operators-z22h2\" (UID: \"62126af2-4cc0-4ccd-bb45-2520abe0f53d\") " pod="openshift-marketplace/redhat-operators-z22h2" Oct 03 13:17:55 crc kubenswrapper[4578]: I1003 13:17:55.146748 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z22h2" Oct 03 13:17:55 crc kubenswrapper[4578]: I1003 13:17:55.299226 4578 generic.go:334] "Generic (PLEG): container finished" podID="6d166384-568f-42ea-a441-4d1df54fd5ce" containerID="c7ea4f5e7503367244dca159b9eacb78bdd9330bfe70bca45fa430ab30564834" exitCode=0 Oct 03 13:17:55 crc kubenswrapper[4578]: I1003 13:17:55.299406 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" event={"ID":"6d166384-568f-42ea-a441-4d1df54fd5ce","Type":"ContainerDied","Data":"c7ea4f5e7503367244dca159b9eacb78bdd9330bfe70bca45fa430ab30564834"} Oct 03 13:17:55 crc kubenswrapper[4578]: I1003 13:17:55.617414 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z22h2"] Oct 03 13:17:55 crc kubenswrapper[4578]: W1003 13:17:55.623916 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62126af2_4cc0_4ccd_bb45_2520abe0f53d.slice/crio-6f942d5f50fb2e1ad159c6eda0c72c2804e86b4db3b61a7db2e8cd3b3bedbe7a WatchSource:0}: Error finding container 6f942d5f50fb2e1ad159c6eda0c72c2804e86b4db3b61a7db2e8cd3b3bedbe7a: Status 404 returned error can't find the container with id 6f942d5f50fb2e1ad159c6eda0c72c2804e86b4db3b61a7db2e8cd3b3bedbe7a Oct 03 13:17:56 crc kubenswrapper[4578]: I1003 13:17:56.308824 4578 generic.go:334] "Generic (PLEG): container finished" podID="62126af2-4cc0-4ccd-bb45-2520abe0f53d" containerID="1a85267d72e8ced72ea115a0412c165809fc62a0ce33c1550dc792b26b28a616" exitCode=0 Oct 03 13:17:56 crc kubenswrapper[4578]: I1003 13:17:56.308923 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z22h2" event={"ID":"62126af2-4cc0-4ccd-bb45-2520abe0f53d","Type":"ContainerDied","Data":"1a85267d72e8ced72ea115a0412c165809fc62a0ce33c1550dc792b26b28a616"} Oct 03 13:17:56 crc kubenswrapper[4578]: I1003 13:17:56.309112 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z22h2" event={"ID":"62126af2-4cc0-4ccd-bb45-2520abe0f53d","Type":"ContainerStarted","Data":"6f942d5f50fb2e1ad159c6eda0c72c2804e86b4db3b61a7db2e8cd3b3bedbe7a"} Oct 03 13:17:56 crc kubenswrapper[4578]: I1003 13:17:56.732866 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" Oct 03 13:17:56 crc kubenswrapper[4578]: I1003 13:17:56.767864 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgcqk\" (UniqueName: \"kubernetes.io/projected/6d166384-568f-42ea-a441-4d1df54fd5ce-kube-api-access-vgcqk\") pod \"6d166384-568f-42ea-a441-4d1df54fd5ce\" (UID: \"6d166384-568f-42ea-a441-4d1df54fd5ce\") " Oct 03 13:17:56 crc kubenswrapper[4578]: I1003 13:17:56.767922 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d166384-568f-42ea-a441-4d1df54fd5ce-inventory\") pod \"6d166384-568f-42ea-a441-4d1df54fd5ce\" (UID: \"6d166384-568f-42ea-a441-4d1df54fd5ce\") " Oct 03 13:17:56 crc kubenswrapper[4578]: I1003 13:17:56.768047 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d166384-568f-42ea-a441-4d1df54fd5ce-repo-setup-combined-ca-bundle\") pod \"6d166384-568f-42ea-a441-4d1df54fd5ce\" (UID: \"6d166384-568f-42ea-a441-4d1df54fd5ce\") " Oct 03 13:17:56 crc kubenswrapper[4578]: I1003 13:17:56.768177 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d166384-568f-42ea-a441-4d1df54fd5ce-ssh-key\") pod \"6d166384-568f-42ea-a441-4d1df54fd5ce\" (UID: \"6d166384-568f-42ea-a441-4d1df54fd5ce\") " Oct 03 13:17:56 crc kubenswrapper[4578]: I1003 13:17:56.777874 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d166384-568f-42ea-a441-4d1df54fd5ce-kube-api-access-vgcqk" (OuterVolumeSpecName: "kube-api-access-vgcqk") pod "6d166384-568f-42ea-a441-4d1df54fd5ce" (UID: "6d166384-568f-42ea-a441-4d1df54fd5ce"). InnerVolumeSpecName "kube-api-access-vgcqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:17:56 crc kubenswrapper[4578]: I1003 13:17:56.785751 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d166384-568f-42ea-a441-4d1df54fd5ce-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "6d166384-568f-42ea-a441-4d1df54fd5ce" (UID: "6d166384-568f-42ea-a441-4d1df54fd5ce"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:17:56 crc kubenswrapper[4578]: I1003 13:17:56.796079 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d166384-568f-42ea-a441-4d1df54fd5ce-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6d166384-568f-42ea-a441-4d1df54fd5ce" (UID: "6d166384-568f-42ea-a441-4d1df54fd5ce"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:17:56 crc kubenswrapper[4578]: I1003 13:17:56.799637 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d166384-568f-42ea-a441-4d1df54fd5ce-inventory" (OuterVolumeSpecName: "inventory") pod "6d166384-568f-42ea-a441-4d1df54fd5ce" (UID: "6d166384-568f-42ea-a441-4d1df54fd5ce"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:17:56 crc kubenswrapper[4578]: I1003 13:17:56.870297 4578 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d166384-568f-42ea-a441-4d1df54fd5ce-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:17:56 crc kubenswrapper[4578]: I1003 13:17:56.870338 4578 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d166384-568f-42ea-a441-4d1df54fd5ce-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:17:56 crc kubenswrapper[4578]: I1003 13:17:56.870351 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgcqk\" (UniqueName: \"kubernetes.io/projected/6d166384-568f-42ea-a441-4d1df54fd5ce-kube-api-access-vgcqk\") on node \"crc\" DevicePath \"\"" Oct 03 13:17:56 crc kubenswrapper[4578]: I1003 13:17:56.870366 4578 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d166384-568f-42ea-a441-4d1df54fd5ce-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.319085 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" event={"ID":"6d166384-568f-42ea-a441-4d1df54fd5ce","Type":"ContainerDied","Data":"62b9a58aa21f584067d1c96d0a6b81becbab23a424a72e7e289318cba48d3d6e"} Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.319127 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62b9a58aa21f584067d1c96d0a6b81becbab23a424a72e7e289318cba48d3d6e" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.319188 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-457zq" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.416348 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x"] Oct 03 13:17:57 crc kubenswrapper[4578]: E1003 13:17:57.417300 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d166384-568f-42ea-a441-4d1df54fd5ce" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.417418 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d166384-568f-42ea-a441-4d1df54fd5ce" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.417755 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d166384-568f-42ea-a441-4d1df54fd5ce" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.418574 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.423579 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.423762 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-442mn" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.423889 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.426586 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.434193 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x"] Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.480595 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a12dea99-870e-4d7b-81e2-f659f1a3c0ee-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ht44x\" (UID: \"a12dea99-870e-4d7b-81e2-f659f1a3c0ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.480678 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a12dea99-870e-4d7b-81e2-f659f1a3c0ee-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ht44x\" (UID: \"a12dea99-870e-4d7b-81e2-f659f1a3c0ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.480765 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfk6c\" (UniqueName: \"kubernetes.io/projected/a12dea99-870e-4d7b-81e2-f659f1a3c0ee-kube-api-access-mfk6c\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ht44x\" (UID: \"a12dea99-870e-4d7b-81e2-f659f1a3c0ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.582613 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a12dea99-870e-4d7b-81e2-f659f1a3c0ee-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ht44x\" (UID: \"a12dea99-870e-4d7b-81e2-f659f1a3c0ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.582938 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a12dea99-870e-4d7b-81e2-f659f1a3c0ee-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ht44x\" (UID: \"a12dea99-870e-4d7b-81e2-f659f1a3c0ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.583021 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfk6c\" (UniqueName: \"kubernetes.io/projected/a12dea99-870e-4d7b-81e2-f659f1a3c0ee-kube-api-access-mfk6c\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ht44x\" (UID: \"a12dea99-870e-4d7b-81e2-f659f1a3c0ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.587812 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a12dea99-870e-4d7b-81e2-f659f1a3c0ee-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ht44x\" (UID: \"a12dea99-870e-4d7b-81e2-f659f1a3c0ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.587829 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a12dea99-870e-4d7b-81e2-f659f1a3c0ee-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ht44x\" (UID: \"a12dea99-870e-4d7b-81e2-f659f1a3c0ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.599394 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfk6c\" (UniqueName: \"kubernetes.io/projected/a12dea99-870e-4d7b-81e2-f659f1a3c0ee-kube-api-access-mfk6c\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ht44x\" (UID: \"a12dea99-870e-4d7b-81e2-f659f1a3c0ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.739060 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x" Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.985017 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-75sqt"] Oct 03 13:17:57 crc kubenswrapper[4578]: I1003 13:17:57.987267 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-75sqt" Oct 03 13:17:58 crc kubenswrapper[4578]: I1003 13:17:58.005097 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-75sqt"] Oct 03 13:17:58 crc kubenswrapper[4578]: I1003 13:17:58.194053 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6qqd\" (UniqueName: \"kubernetes.io/projected/31df414b-9e4d-4a89-be61-17c99820ce42-kube-api-access-h6qqd\") pod \"redhat-marketplace-75sqt\" (UID: \"31df414b-9e4d-4a89-be61-17c99820ce42\") " pod="openshift-marketplace/redhat-marketplace-75sqt" Oct 03 13:17:58 crc kubenswrapper[4578]: I1003 13:17:58.194156 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31df414b-9e4d-4a89-be61-17c99820ce42-catalog-content\") pod \"redhat-marketplace-75sqt\" (UID: \"31df414b-9e4d-4a89-be61-17c99820ce42\") " pod="openshift-marketplace/redhat-marketplace-75sqt" Oct 03 13:17:58 crc kubenswrapper[4578]: I1003 13:17:58.194201 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31df414b-9e4d-4a89-be61-17c99820ce42-utilities\") pod \"redhat-marketplace-75sqt\" (UID: \"31df414b-9e4d-4a89-be61-17c99820ce42\") " pod="openshift-marketplace/redhat-marketplace-75sqt" Oct 03 13:17:58 crc kubenswrapper[4578]: I1003 13:17:58.274677 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x"] Oct 03 13:17:58 crc kubenswrapper[4578]: I1003 13:17:58.295647 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31df414b-9e4d-4a89-be61-17c99820ce42-utilities\") pod \"redhat-marketplace-75sqt\" (UID: \"31df414b-9e4d-4a89-be61-17c99820ce42\") " pod="openshift-marketplace/redhat-marketplace-75sqt" Oct 03 13:17:58 crc kubenswrapper[4578]: I1003 13:17:58.295776 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6qqd\" (UniqueName: \"kubernetes.io/projected/31df414b-9e4d-4a89-be61-17c99820ce42-kube-api-access-h6qqd\") pod \"redhat-marketplace-75sqt\" (UID: \"31df414b-9e4d-4a89-be61-17c99820ce42\") " pod="openshift-marketplace/redhat-marketplace-75sqt" Oct 03 13:17:58 crc kubenswrapper[4578]: I1003 13:17:58.295834 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31df414b-9e4d-4a89-be61-17c99820ce42-catalog-content\") pod \"redhat-marketplace-75sqt\" (UID: \"31df414b-9e4d-4a89-be61-17c99820ce42\") " pod="openshift-marketplace/redhat-marketplace-75sqt" Oct 03 13:17:58 crc kubenswrapper[4578]: I1003 13:17:58.296917 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31df414b-9e4d-4a89-be61-17c99820ce42-utilities\") pod \"redhat-marketplace-75sqt\" (UID: \"31df414b-9e4d-4a89-be61-17c99820ce42\") " pod="openshift-marketplace/redhat-marketplace-75sqt" Oct 03 13:17:58 crc kubenswrapper[4578]: I1003 13:17:58.297063 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31df414b-9e4d-4a89-be61-17c99820ce42-catalog-content\") pod \"redhat-marketplace-75sqt\" (UID: \"31df414b-9e4d-4a89-be61-17c99820ce42\") " pod="openshift-marketplace/redhat-marketplace-75sqt" Oct 03 13:17:58 crc kubenswrapper[4578]: I1003 13:17:58.321135 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6qqd\" (UniqueName: \"kubernetes.io/projected/31df414b-9e4d-4a89-be61-17c99820ce42-kube-api-access-h6qqd\") pod \"redhat-marketplace-75sqt\" (UID: \"31df414b-9e4d-4a89-be61-17c99820ce42\") " pod="openshift-marketplace/redhat-marketplace-75sqt" Oct 03 13:17:58 crc kubenswrapper[4578]: I1003 13:17:58.329833 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-75sqt" Oct 03 13:17:58 crc kubenswrapper[4578]: I1003 13:17:58.332918 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z22h2" event={"ID":"62126af2-4cc0-4ccd-bb45-2520abe0f53d","Type":"ContainerStarted","Data":"b260c00dce59b57307f9c4888c07d246b7f1c0bbd7846bff72e14ea476511b4b"} Oct 03 13:17:58 crc kubenswrapper[4578]: I1003 13:17:58.334723 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x" event={"ID":"a12dea99-870e-4d7b-81e2-f659f1a3c0ee","Type":"ContainerStarted","Data":"13b5edf28166bc035213befc6c5f1ea9eb047188ebec99ac4cdf18d2296fe599"} Oct 03 13:17:58 crc kubenswrapper[4578]: I1003 13:17:58.793132 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-75sqt"] Oct 03 13:17:59 crc kubenswrapper[4578]: I1003 13:17:59.346024 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x" event={"ID":"a12dea99-870e-4d7b-81e2-f659f1a3c0ee","Type":"ContainerStarted","Data":"ebbed7025ada2f8200b024e4dbb89056bf2a52cf5982086b8f08e4475127610b"} Oct 03 13:17:59 crc kubenswrapper[4578]: I1003 13:17:59.349660 4578 generic.go:334] "Generic (PLEG): container finished" podID="31df414b-9e4d-4a89-be61-17c99820ce42" containerID="7b45833c8ed76e2d85d86dba15e26dc8b4e08595b8805c24a00e46e44122a37d" exitCode=0 Oct 03 13:17:59 crc kubenswrapper[4578]: I1003 13:17:59.350578 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-75sqt" event={"ID":"31df414b-9e4d-4a89-be61-17c99820ce42","Type":"ContainerDied","Data":"7b45833c8ed76e2d85d86dba15e26dc8b4e08595b8805c24a00e46e44122a37d"} Oct 03 13:17:59 crc kubenswrapper[4578]: I1003 13:17:59.350618 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-75sqt" event={"ID":"31df414b-9e4d-4a89-be61-17c99820ce42","Type":"ContainerStarted","Data":"2367cde42549e5466d78adbfa4f1e9fe7d8d907d8400ee77139f1e5d277fe195"} Oct 03 13:17:59 crc kubenswrapper[4578]: I1003 13:17:59.387766 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x" podStartSLOduration=2.1553910370000002 podStartE2EDuration="2.387742706s" podCreationTimestamp="2025-10-03 13:17:57 +0000 UTC" firstStartedPulling="2025-10-03 13:17:58.286842812 +0000 UTC m=+1614.085314996" lastFinishedPulling="2025-10-03 13:17:58.519194481 +0000 UTC m=+1614.317666665" observedRunningTime="2025-10-03 13:17:59.367895063 +0000 UTC m=+1615.166367247" watchObservedRunningTime="2025-10-03 13:17:59.387742706 +0000 UTC m=+1615.186214890" Oct 03 13:18:00 crc kubenswrapper[4578]: I1003 13:18:00.361888 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-75sqt" event={"ID":"31df414b-9e4d-4a89-be61-17c99820ce42","Type":"ContainerStarted","Data":"284cc49070c42c1df5cb6e6ced583280df4a17c13f3e56c8fb805dae3bdc35ee"} Oct 03 13:18:02 crc kubenswrapper[4578]: I1003 13:18:02.382169 4578 generic.go:334] "Generic (PLEG): container finished" podID="a12dea99-870e-4d7b-81e2-f659f1a3c0ee" containerID="ebbed7025ada2f8200b024e4dbb89056bf2a52cf5982086b8f08e4475127610b" exitCode=0 Oct 03 13:18:02 crc kubenswrapper[4578]: I1003 13:18:02.382612 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x" event={"ID":"a12dea99-870e-4d7b-81e2-f659f1a3c0ee","Type":"ContainerDied","Data":"ebbed7025ada2f8200b024e4dbb89056bf2a52cf5982086b8f08e4475127610b"} Oct 03 13:18:02 crc kubenswrapper[4578]: I1003 13:18:02.386095 4578 generic.go:334] "Generic (PLEG): container finished" podID="62126af2-4cc0-4ccd-bb45-2520abe0f53d" containerID="b260c00dce59b57307f9c4888c07d246b7f1c0bbd7846bff72e14ea476511b4b" exitCode=0 Oct 03 13:18:02 crc kubenswrapper[4578]: I1003 13:18:02.386145 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z22h2" event={"ID":"62126af2-4cc0-4ccd-bb45-2520abe0f53d","Type":"ContainerDied","Data":"b260c00dce59b57307f9c4888c07d246b7f1c0bbd7846bff72e14ea476511b4b"} Oct 03 13:18:02 crc kubenswrapper[4578]: I1003 13:18:02.390488 4578 generic.go:334] "Generic (PLEG): container finished" podID="31df414b-9e4d-4a89-be61-17c99820ce42" containerID="284cc49070c42c1df5cb6e6ced583280df4a17c13f3e56c8fb805dae3bdc35ee" exitCode=0 Oct 03 13:18:02 crc kubenswrapper[4578]: I1003 13:18:02.390517 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-75sqt" event={"ID":"31df414b-9e4d-4a89-be61-17c99820ce42","Type":"ContainerDied","Data":"284cc49070c42c1df5cb6e6ced583280df4a17c13f3e56c8fb805dae3bdc35ee"} Oct 03 13:18:03 crc kubenswrapper[4578]: I1003 13:18:03.421437 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z22h2" event={"ID":"62126af2-4cc0-4ccd-bb45-2520abe0f53d","Type":"ContainerStarted","Data":"111a59d9921136e8b11e164a00414d355443a33f47a7537387c0fbeaa068caf6"} Oct 03 13:18:03 crc kubenswrapper[4578]: I1003 13:18:03.427123 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-75sqt" event={"ID":"31df414b-9e4d-4a89-be61-17c99820ce42","Type":"ContainerStarted","Data":"b23503a5926daa2c5161c5668e5bb490fcd0d51d09440a0348683cd0a4ce0d8a"} Oct 03 13:18:03 crc kubenswrapper[4578]: I1003 13:18:03.442264 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z22h2" podStartSLOduration=2.588644362 podStartE2EDuration="9.442243858s" podCreationTimestamp="2025-10-03 13:17:54 +0000 UTC" firstStartedPulling="2025-10-03 13:17:56.311258078 +0000 UTC m=+1612.109730262" lastFinishedPulling="2025-10-03 13:18:03.164857574 +0000 UTC m=+1618.963329758" observedRunningTime="2025-10-03 13:18:03.437516789 +0000 UTC m=+1619.235988983" watchObservedRunningTime="2025-10-03 13:18:03.442243858 +0000 UTC m=+1619.240716042" Oct 03 13:18:03 crc kubenswrapper[4578]: I1003 13:18:03.473994 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-75sqt" podStartSLOduration=2.742188282 podStartE2EDuration="6.473973105s" podCreationTimestamp="2025-10-03 13:17:57 +0000 UTC" firstStartedPulling="2025-10-03 13:17:59.35188764 +0000 UTC m=+1615.150359824" lastFinishedPulling="2025-10-03 13:18:03.083672463 +0000 UTC m=+1618.882144647" observedRunningTime="2025-10-03 13:18:03.464470456 +0000 UTC m=+1619.262942650" watchObservedRunningTime="2025-10-03 13:18:03.473973105 +0000 UTC m=+1619.272445299" Oct 03 13:18:03 crc kubenswrapper[4578]: I1003 13:18:03.856907 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x" Oct 03 13:18:03 crc kubenswrapper[4578]: I1003 13:18:03.908917 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:18:03 crc kubenswrapper[4578]: E1003 13:18:03.909187 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.003253 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a12dea99-870e-4d7b-81e2-f659f1a3c0ee-inventory\") pod \"a12dea99-870e-4d7b-81e2-f659f1a3c0ee\" (UID: \"a12dea99-870e-4d7b-81e2-f659f1a3c0ee\") " Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.003480 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfk6c\" (UniqueName: \"kubernetes.io/projected/a12dea99-870e-4d7b-81e2-f659f1a3c0ee-kube-api-access-mfk6c\") pod \"a12dea99-870e-4d7b-81e2-f659f1a3c0ee\" (UID: \"a12dea99-870e-4d7b-81e2-f659f1a3c0ee\") " Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.003597 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a12dea99-870e-4d7b-81e2-f659f1a3c0ee-ssh-key\") pod \"a12dea99-870e-4d7b-81e2-f659f1a3c0ee\" (UID: \"a12dea99-870e-4d7b-81e2-f659f1a3c0ee\") " Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.009159 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a12dea99-870e-4d7b-81e2-f659f1a3c0ee-kube-api-access-mfk6c" (OuterVolumeSpecName: "kube-api-access-mfk6c") pod "a12dea99-870e-4d7b-81e2-f659f1a3c0ee" (UID: "a12dea99-870e-4d7b-81e2-f659f1a3c0ee"). InnerVolumeSpecName "kube-api-access-mfk6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.033031 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a12dea99-870e-4d7b-81e2-f659f1a3c0ee-inventory" (OuterVolumeSpecName: "inventory") pod "a12dea99-870e-4d7b-81e2-f659f1a3c0ee" (UID: "a12dea99-870e-4d7b-81e2-f659f1a3c0ee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.035981 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a12dea99-870e-4d7b-81e2-f659f1a3c0ee-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a12dea99-870e-4d7b-81e2-f659f1a3c0ee" (UID: "a12dea99-870e-4d7b-81e2-f659f1a3c0ee"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.105322 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfk6c\" (UniqueName: \"kubernetes.io/projected/a12dea99-870e-4d7b-81e2-f659f1a3c0ee-kube-api-access-mfk6c\") on node \"crc\" DevicePath \"\"" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.105355 4578 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a12dea99-870e-4d7b-81e2-f659f1a3c0ee-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.105367 4578 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a12dea99-870e-4d7b-81e2-f659f1a3c0ee-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.437245 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x" event={"ID":"a12dea99-870e-4d7b-81e2-f659f1a3c0ee","Type":"ContainerDied","Data":"13b5edf28166bc035213befc6c5f1ea9eb047188ebec99ac4cdf18d2296fe599"} Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.437281 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13b5edf28166bc035213befc6c5f1ea9eb047188ebec99ac4cdf18d2296fe599" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.437334 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ht44x" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.482075 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8"] Oct 03 13:18:04 crc kubenswrapper[4578]: E1003 13:18:04.482519 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a12dea99-870e-4d7b-81e2-f659f1a3c0ee" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.482537 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a12dea99-870e-4d7b-81e2-f659f1a3c0ee" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.483060 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="a12dea99-870e-4d7b-81e2-f659f1a3c0ee" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.483696 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.487131 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.487132 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-442mn" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.487180 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.487545 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.533310 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8"] Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.615327 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w58gn\" (UniqueName: \"kubernetes.io/projected/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-kube-api-access-w58gn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8\" (UID: \"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.615457 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8\" (UID: \"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.615557 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8\" (UID: \"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.615642 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8\" (UID: \"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.717795 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8\" (UID: \"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.717925 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8\" (UID: \"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.717962 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w58gn\" (UniqueName: \"kubernetes.io/projected/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-kube-api-access-w58gn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8\" (UID: \"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.718002 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8\" (UID: \"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.721956 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8\" (UID: \"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.722321 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8\" (UID: \"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.723492 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8\" (UID: \"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.742660 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w58gn\" (UniqueName: \"kubernetes.io/projected/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-kube-api-access-w58gn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8\" (UID: \"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" Oct 03 13:18:04 crc kubenswrapper[4578]: I1003 13:18:04.810425 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" Oct 03 13:18:05 crc kubenswrapper[4578]: I1003 13:18:05.148017 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z22h2" Oct 03 13:18:05 crc kubenswrapper[4578]: I1003 13:18:05.148396 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z22h2" Oct 03 13:18:05 crc kubenswrapper[4578]: I1003 13:18:05.403170 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8"] Oct 03 13:18:05 crc kubenswrapper[4578]: I1003 13:18:05.448571 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" event={"ID":"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4","Type":"ContainerStarted","Data":"b27a7f7137642c2e3e6f5717c103e202f0ab9cc254b2f882ef523cfe928fb7f5"} Oct 03 13:18:05 crc kubenswrapper[4578]: I1003 13:18:05.577727 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:18:06 crc kubenswrapper[4578]: I1003 13:18:06.200470 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-z22h2" podUID="62126af2-4cc0-4ccd-bb45-2520abe0f53d" containerName="registry-server" probeResult="failure" output=< Oct 03 13:18:06 crc kubenswrapper[4578]: timeout: failed to connect service ":50051" within 1s Oct 03 13:18:06 crc kubenswrapper[4578]: > Oct 03 13:18:06 crc kubenswrapper[4578]: I1003 13:18:06.461052 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" event={"ID":"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4","Type":"ContainerStarted","Data":"f86d2b3f43591a95ad325feb387f015bf4621162b162c0e2724af06df0ca54e1"} Oct 03 13:18:06 crc kubenswrapper[4578]: I1003 13:18:06.485935 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" podStartSLOduration=2.313562911 podStartE2EDuration="2.485919945s" podCreationTimestamp="2025-10-03 13:18:04 +0000 UTC" firstStartedPulling="2025-10-03 13:18:05.402339124 +0000 UTC m=+1621.200811298" lastFinishedPulling="2025-10-03 13:18:05.574696148 +0000 UTC m=+1621.373168332" observedRunningTime="2025-10-03 13:18:06.475745745 +0000 UTC m=+1622.274217949" watchObservedRunningTime="2025-10-03 13:18:06.485919945 +0000 UTC m=+1622.284392129" Oct 03 13:18:07 crc kubenswrapper[4578]: I1003 13:18:07.146267 4578 scope.go:117] "RemoveContainer" containerID="908f345453ffb6da3b63f9535168cdc1bdc29491e2e16489faccc06fb38326f0" Oct 03 13:18:08 crc kubenswrapper[4578]: I1003 13:18:08.331079 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-75sqt" Oct 03 13:18:08 crc kubenswrapper[4578]: I1003 13:18:08.331507 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-75sqt" Oct 03 13:18:08 crc kubenswrapper[4578]: I1003 13:18:08.379100 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-75sqt" Oct 03 13:18:08 crc kubenswrapper[4578]: I1003 13:18:08.526561 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-75sqt" Oct 03 13:18:08 crc kubenswrapper[4578]: I1003 13:18:08.771407 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-75sqt"] Oct 03 13:18:10 crc kubenswrapper[4578]: I1003 13:18:10.491854 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-75sqt" podUID="31df414b-9e4d-4a89-be61-17c99820ce42" containerName="registry-server" containerID="cri-o://b23503a5926daa2c5161c5668e5bb490fcd0d51d09440a0348683cd0a4ce0d8a" gracePeriod=2 Oct 03 13:18:10 crc kubenswrapper[4578]: I1003 13:18:10.970704 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-75sqt" Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.049178 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31df414b-9e4d-4a89-be61-17c99820ce42-utilities\") pod \"31df414b-9e4d-4a89-be61-17c99820ce42\" (UID: \"31df414b-9e4d-4a89-be61-17c99820ce42\") " Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.049264 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6qqd\" (UniqueName: \"kubernetes.io/projected/31df414b-9e4d-4a89-be61-17c99820ce42-kube-api-access-h6qqd\") pod \"31df414b-9e4d-4a89-be61-17c99820ce42\" (UID: \"31df414b-9e4d-4a89-be61-17c99820ce42\") " Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.049333 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31df414b-9e4d-4a89-be61-17c99820ce42-catalog-content\") pod \"31df414b-9e4d-4a89-be61-17c99820ce42\" (UID: \"31df414b-9e4d-4a89-be61-17c99820ce42\") " Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.049959 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31df414b-9e4d-4a89-be61-17c99820ce42-utilities" (OuterVolumeSpecName: "utilities") pod "31df414b-9e4d-4a89-be61-17c99820ce42" (UID: "31df414b-9e4d-4a89-be61-17c99820ce42"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.055736 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31df414b-9e4d-4a89-be61-17c99820ce42-kube-api-access-h6qqd" (OuterVolumeSpecName: "kube-api-access-h6qqd") pod "31df414b-9e4d-4a89-be61-17c99820ce42" (UID: "31df414b-9e4d-4a89-be61-17c99820ce42"). InnerVolumeSpecName "kube-api-access-h6qqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.064378 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31df414b-9e4d-4a89-be61-17c99820ce42-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "31df414b-9e4d-4a89-be61-17c99820ce42" (UID: "31df414b-9e4d-4a89-be61-17c99820ce42"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.151061 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31df414b-9e4d-4a89-be61-17c99820ce42-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.151094 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6qqd\" (UniqueName: \"kubernetes.io/projected/31df414b-9e4d-4a89-be61-17c99820ce42-kube-api-access-h6qqd\") on node \"crc\" DevicePath \"\"" Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.151106 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31df414b-9e4d-4a89-be61-17c99820ce42-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.503311 4578 generic.go:334] "Generic (PLEG): container finished" podID="31df414b-9e4d-4a89-be61-17c99820ce42" containerID="b23503a5926daa2c5161c5668e5bb490fcd0d51d09440a0348683cd0a4ce0d8a" exitCode=0 Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.503360 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-75sqt" Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.503378 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-75sqt" event={"ID":"31df414b-9e4d-4a89-be61-17c99820ce42","Type":"ContainerDied","Data":"b23503a5926daa2c5161c5668e5bb490fcd0d51d09440a0348683cd0a4ce0d8a"} Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.503764 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-75sqt" event={"ID":"31df414b-9e4d-4a89-be61-17c99820ce42","Type":"ContainerDied","Data":"2367cde42549e5466d78adbfa4f1e9fe7d8d907d8400ee77139f1e5d277fe195"} Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.503785 4578 scope.go:117] "RemoveContainer" containerID="b23503a5926daa2c5161c5668e5bb490fcd0d51d09440a0348683cd0a4ce0d8a" Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.524200 4578 scope.go:117] "RemoveContainer" containerID="284cc49070c42c1df5cb6e6ced583280df4a17c13f3e56c8fb805dae3bdc35ee" Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.562728 4578 scope.go:117] "RemoveContainer" containerID="7b45833c8ed76e2d85d86dba15e26dc8b4e08595b8805c24a00e46e44122a37d" Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.564343 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-75sqt"] Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.576792 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-75sqt"] Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.637555 4578 scope.go:117] "RemoveContainer" containerID="b23503a5926daa2c5161c5668e5bb490fcd0d51d09440a0348683cd0a4ce0d8a" Oct 03 13:18:11 crc kubenswrapper[4578]: E1003 13:18:11.638964 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b23503a5926daa2c5161c5668e5bb490fcd0d51d09440a0348683cd0a4ce0d8a\": container with ID starting with b23503a5926daa2c5161c5668e5bb490fcd0d51d09440a0348683cd0a4ce0d8a not found: ID does not exist" containerID="b23503a5926daa2c5161c5668e5bb490fcd0d51d09440a0348683cd0a4ce0d8a" Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.639010 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b23503a5926daa2c5161c5668e5bb490fcd0d51d09440a0348683cd0a4ce0d8a"} err="failed to get container status \"b23503a5926daa2c5161c5668e5bb490fcd0d51d09440a0348683cd0a4ce0d8a\": rpc error: code = NotFound desc = could not find container \"b23503a5926daa2c5161c5668e5bb490fcd0d51d09440a0348683cd0a4ce0d8a\": container with ID starting with b23503a5926daa2c5161c5668e5bb490fcd0d51d09440a0348683cd0a4ce0d8a not found: ID does not exist" Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.639037 4578 scope.go:117] "RemoveContainer" containerID="284cc49070c42c1df5cb6e6ced583280df4a17c13f3e56c8fb805dae3bdc35ee" Oct 03 13:18:11 crc kubenswrapper[4578]: E1003 13:18:11.639501 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"284cc49070c42c1df5cb6e6ced583280df4a17c13f3e56c8fb805dae3bdc35ee\": container with ID starting with 284cc49070c42c1df5cb6e6ced583280df4a17c13f3e56c8fb805dae3bdc35ee not found: ID does not exist" containerID="284cc49070c42c1df5cb6e6ced583280df4a17c13f3e56c8fb805dae3bdc35ee" Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.639533 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"284cc49070c42c1df5cb6e6ced583280df4a17c13f3e56c8fb805dae3bdc35ee"} err="failed to get container status \"284cc49070c42c1df5cb6e6ced583280df4a17c13f3e56c8fb805dae3bdc35ee\": rpc error: code = NotFound desc = could not find container \"284cc49070c42c1df5cb6e6ced583280df4a17c13f3e56c8fb805dae3bdc35ee\": container with ID starting with 284cc49070c42c1df5cb6e6ced583280df4a17c13f3e56c8fb805dae3bdc35ee not found: ID does not exist" Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.639581 4578 scope.go:117] "RemoveContainer" containerID="7b45833c8ed76e2d85d86dba15e26dc8b4e08595b8805c24a00e46e44122a37d" Oct 03 13:18:11 crc kubenswrapper[4578]: E1003 13:18:11.639921 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b45833c8ed76e2d85d86dba15e26dc8b4e08595b8805c24a00e46e44122a37d\": container with ID starting with 7b45833c8ed76e2d85d86dba15e26dc8b4e08595b8805c24a00e46e44122a37d not found: ID does not exist" containerID="7b45833c8ed76e2d85d86dba15e26dc8b4e08595b8805c24a00e46e44122a37d" Oct 03 13:18:11 crc kubenswrapper[4578]: I1003 13:18:11.639945 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b45833c8ed76e2d85d86dba15e26dc8b4e08595b8805c24a00e46e44122a37d"} err="failed to get container status \"7b45833c8ed76e2d85d86dba15e26dc8b4e08595b8805c24a00e46e44122a37d\": rpc error: code = NotFound desc = could not find container \"7b45833c8ed76e2d85d86dba15e26dc8b4e08595b8805c24a00e46e44122a37d\": container with ID starting with 7b45833c8ed76e2d85d86dba15e26dc8b4e08595b8805c24a00e46e44122a37d not found: ID does not exist" Oct 03 13:18:12 crc kubenswrapper[4578]: I1003 13:18:12.921062 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31df414b-9e4d-4a89-be61-17c99820ce42" path="/var/lib/kubelet/pods/31df414b-9e4d-4a89-be61-17c99820ce42/volumes" Oct 03 13:18:16 crc kubenswrapper[4578]: I1003 13:18:16.194919 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-z22h2" podUID="62126af2-4cc0-4ccd-bb45-2520abe0f53d" containerName="registry-server" probeResult="failure" output=< Oct 03 13:18:16 crc kubenswrapper[4578]: timeout: failed to connect service ":50051" within 1s Oct 03 13:18:16 crc kubenswrapper[4578]: > Oct 03 13:18:16 crc kubenswrapper[4578]: I1003 13:18:16.909328 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:18:16 crc kubenswrapper[4578]: E1003 13:18:16.909650 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:18:25 crc kubenswrapper[4578]: I1003 13:18:25.197321 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z22h2" Oct 03 13:18:25 crc kubenswrapper[4578]: I1003 13:18:25.245686 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z22h2" Oct 03 13:18:25 crc kubenswrapper[4578]: I1003 13:18:25.982602 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z22h2"] Oct 03 13:18:26 crc kubenswrapper[4578]: I1003 13:18:26.640515 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z22h2" podUID="62126af2-4cc0-4ccd-bb45-2520abe0f53d" containerName="registry-server" containerID="cri-o://111a59d9921136e8b11e164a00414d355443a33f47a7537387c0fbeaa068caf6" gracePeriod=2 Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.046709 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z22h2" Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.228240 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62126af2-4cc0-4ccd-bb45-2520abe0f53d-utilities\") pod \"62126af2-4cc0-4ccd-bb45-2520abe0f53d\" (UID: \"62126af2-4cc0-4ccd-bb45-2520abe0f53d\") " Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.228365 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xqp6\" (UniqueName: \"kubernetes.io/projected/62126af2-4cc0-4ccd-bb45-2520abe0f53d-kube-api-access-9xqp6\") pod \"62126af2-4cc0-4ccd-bb45-2520abe0f53d\" (UID: \"62126af2-4cc0-4ccd-bb45-2520abe0f53d\") " Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.228533 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62126af2-4cc0-4ccd-bb45-2520abe0f53d-catalog-content\") pod \"62126af2-4cc0-4ccd-bb45-2520abe0f53d\" (UID: \"62126af2-4cc0-4ccd-bb45-2520abe0f53d\") " Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.229002 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62126af2-4cc0-4ccd-bb45-2520abe0f53d-utilities" (OuterVolumeSpecName: "utilities") pod "62126af2-4cc0-4ccd-bb45-2520abe0f53d" (UID: "62126af2-4cc0-4ccd-bb45-2520abe0f53d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.239481 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62126af2-4cc0-4ccd-bb45-2520abe0f53d-kube-api-access-9xqp6" (OuterVolumeSpecName: "kube-api-access-9xqp6") pod "62126af2-4cc0-4ccd-bb45-2520abe0f53d" (UID: "62126af2-4cc0-4ccd-bb45-2520abe0f53d"). InnerVolumeSpecName "kube-api-access-9xqp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.315963 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62126af2-4cc0-4ccd-bb45-2520abe0f53d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "62126af2-4cc0-4ccd-bb45-2520abe0f53d" (UID: "62126af2-4cc0-4ccd-bb45-2520abe0f53d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.330959 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62126af2-4cc0-4ccd-bb45-2520abe0f53d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.330987 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62126af2-4cc0-4ccd-bb45-2520abe0f53d-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.330997 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xqp6\" (UniqueName: \"kubernetes.io/projected/62126af2-4cc0-4ccd-bb45-2520abe0f53d-kube-api-access-9xqp6\") on node \"crc\" DevicePath \"\"" Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.651993 4578 generic.go:334] "Generic (PLEG): container finished" podID="62126af2-4cc0-4ccd-bb45-2520abe0f53d" containerID="111a59d9921136e8b11e164a00414d355443a33f47a7537387c0fbeaa068caf6" exitCode=0 Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.652063 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z22h2" Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.652052 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z22h2" event={"ID":"62126af2-4cc0-4ccd-bb45-2520abe0f53d","Type":"ContainerDied","Data":"111a59d9921136e8b11e164a00414d355443a33f47a7537387c0fbeaa068caf6"} Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.652220 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z22h2" event={"ID":"62126af2-4cc0-4ccd-bb45-2520abe0f53d","Type":"ContainerDied","Data":"6f942d5f50fb2e1ad159c6eda0c72c2804e86b4db3b61a7db2e8cd3b3bedbe7a"} Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.652244 4578 scope.go:117] "RemoveContainer" containerID="111a59d9921136e8b11e164a00414d355443a33f47a7537387c0fbeaa068caf6" Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.689094 4578 scope.go:117] "RemoveContainer" containerID="b260c00dce59b57307f9c4888c07d246b7f1c0bbd7846bff72e14ea476511b4b" Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.690780 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z22h2"] Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.699983 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z22h2"] Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.714241 4578 scope.go:117] "RemoveContainer" containerID="1a85267d72e8ced72ea115a0412c165809fc62a0ce33c1550dc792b26b28a616" Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.757476 4578 scope.go:117] "RemoveContainer" containerID="111a59d9921136e8b11e164a00414d355443a33f47a7537387c0fbeaa068caf6" Oct 03 13:18:27 crc kubenswrapper[4578]: E1003 13:18:27.758547 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"111a59d9921136e8b11e164a00414d355443a33f47a7537387c0fbeaa068caf6\": container with ID starting with 111a59d9921136e8b11e164a00414d355443a33f47a7537387c0fbeaa068caf6 not found: ID does not exist" containerID="111a59d9921136e8b11e164a00414d355443a33f47a7537387c0fbeaa068caf6" Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.758595 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"111a59d9921136e8b11e164a00414d355443a33f47a7537387c0fbeaa068caf6"} err="failed to get container status \"111a59d9921136e8b11e164a00414d355443a33f47a7537387c0fbeaa068caf6\": rpc error: code = NotFound desc = could not find container \"111a59d9921136e8b11e164a00414d355443a33f47a7537387c0fbeaa068caf6\": container with ID starting with 111a59d9921136e8b11e164a00414d355443a33f47a7537387c0fbeaa068caf6 not found: ID does not exist" Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.758953 4578 scope.go:117] "RemoveContainer" containerID="b260c00dce59b57307f9c4888c07d246b7f1c0bbd7846bff72e14ea476511b4b" Oct 03 13:18:27 crc kubenswrapper[4578]: E1003 13:18:27.759676 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b260c00dce59b57307f9c4888c07d246b7f1c0bbd7846bff72e14ea476511b4b\": container with ID starting with b260c00dce59b57307f9c4888c07d246b7f1c0bbd7846bff72e14ea476511b4b not found: ID does not exist" containerID="b260c00dce59b57307f9c4888c07d246b7f1c0bbd7846bff72e14ea476511b4b" Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.759710 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b260c00dce59b57307f9c4888c07d246b7f1c0bbd7846bff72e14ea476511b4b"} err="failed to get container status \"b260c00dce59b57307f9c4888c07d246b7f1c0bbd7846bff72e14ea476511b4b\": rpc error: code = NotFound desc = could not find container \"b260c00dce59b57307f9c4888c07d246b7f1c0bbd7846bff72e14ea476511b4b\": container with ID starting with b260c00dce59b57307f9c4888c07d246b7f1c0bbd7846bff72e14ea476511b4b not found: ID does not exist" Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.759730 4578 scope.go:117] "RemoveContainer" containerID="1a85267d72e8ced72ea115a0412c165809fc62a0ce33c1550dc792b26b28a616" Oct 03 13:18:27 crc kubenswrapper[4578]: E1003 13:18:27.759972 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a85267d72e8ced72ea115a0412c165809fc62a0ce33c1550dc792b26b28a616\": container with ID starting with 1a85267d72e8ced72ea115a0412c165809fc62a0ce33c1550dc792b26b28a616 not found: ID does not exist" containerID="1a85267d72e8ced72ea115a0412c165809fc62a0ce33c1550dc792b26b28a616" Oct 03 13:18:27 crc kubenswrapper[4578]: I1003 13:18:27.759997 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a85267d72e8ced72ea115a0412c165809fc62a0ce33c1550dc792b26b28a616"} err="failed to get container status \"1a85267d72e8ced72ea115a0412c165809fc62a0ce33c1550dc792b26b28a616\": rpc error: code = NotFound desc = could not find container \"1a85267d72e8ced72ea115a0412c165809fc62a0ce33c1550dc792b26b28a616\": container with ID starting with 1a85267d72e8ced72ea115a0412c165809fc62a0ce33c1550dc792b26b28a616 not found: ID does not exist" Oct 03 13:18:28 crc kubenswrapper[4578]: I1003 13:18:28.921522 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62126af2-4cc0-4ccd-bb45-2520abe0f53d" path="/var/lib/kubelet/pods/62126af2-4cc0-4ccd-bb45-2520abe0f53d/volumes" Oct 03 13:18:31 crc kubenswrapper[4578]: I1003 13:18:31.910082 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:18:31 crc kubenswrapper[4578]: E1003 13:18:31.911553 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:18:44 crc kubenswrapper[4578]: I1003 13:18:44.915820 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:18:44 crc kubenswrapper[4578]: E1003 13:18:44.916684 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:18:55 crc kubenswrapper[4578]: I1003 13:18:55.909195 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:18:55 crc kubenswrapper[4578]: E1003 13:18:55.909943 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:19:06 crc kubenswrapper[4578]: I1003 13:19:06.908904 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:19:06 crc kubenswrapper[4578]: E1003 13:19:06.909615 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:19:07 crc kubenswrapper[4578]: I1003 13:19:07.205708 4578 scope.go:117] "RemoveContainer" containerID="054c0e5021dfccfa10b4b49660eacf259766addadd0d456e3e6738b69fd8d4ef" Oct 03 13:19:18 crc kubenswrapper[4578]: I1003 13:19:18.909408 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:19:18 crc kubenswrapper[4578]: E1003 13:19:18.910246 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:19:31 crc kubenswrapper[4578]: I1003 13:19:31.909713 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:19:31 crc kubenswrapper[4578]: E1003 13:19:31.910718 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:19:43 crc kubenswrapper[4578]: I1003 13:19:43.909310 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:19:43 crc kubenswrapper[4578]: E1003 13:19:43.910102 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:19:56 crc kubenswrapper[4578]: I1003 13:19:56.909601 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:19:56 crc kubenswrapper[4578]: E1003 13:19:56.910404 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:20:07 crc kubenswrapper[4578]: I1003 13:20:07.335177 4578 scope.go:117] "RemoveContainer" containerID="6b5fa4fe8900e80130f7c569ae25f05762f0b391172a3db23a44840feff5ad7d" Oct 03 13:20:07 crc kubenswrapper[4578]: I1003 13:20:07.359940 4578 scope.go:117] "RemoveContainer" containerID="8e9ae4feb2c38fbae277c3dbbdc74ea68b47414938aebd6eabba698f4914b621" Oct 03 13:20:07 crc kubenswrapper[4578]: I1003 13:20:07.387918 4578 scope.go:117] "RemoveContainer" containerID="9c2664687a2e5ce8899e53ba9c5dfffce4a6572d01dd5e672b8a756e71a4722a" Oct 03 13:20:07 crc kubenswrapper[4578]: I1003 13:20:07.410336 4578 scope.go:117] "RemoveContainer" containerID="6853ac4b339f037dbd2c66955d65aaca7d497983ca6a0d31ff77dc5e57434c4e" Oct 03 13:20:10 crc kubenswrapper[4578]: I1003 13:20:10.910142 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:20:10 crc kubenswrapper[4578]: E1003 13:20:10.910718 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:20:23 crc kubenswrapper[4578]: I1003 13:20:23.909535 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:20:23 crc kubenswrapper[4578]: E1003 13:20:23.910216 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:20:38 crc kubenswrapper[4578]: I1003 13:20:38.909059 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:20:38 crc kubenswrapper[4578]: E1003 13:20:38.909921 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:20:50 crc kubenswrapper[4578]: I1003 13:20:50.909275 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:20:50 crc kubenswrapper[4578]: E1003 13:20:50.910614 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:21:01 crc kubenswrapper[4578]: I1003 13:21:01.909824 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:21:01 crc kubenswrapper[4578]: E1003 13:21:01.910603 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:21:07 crc kubenswrapper[4578]: I1003 13:21:07.460100 4578 scope.go:117] "RemoveContainer" containerID="9570a50ac3ed7d4fc6eebdda73ed2bba5de9e90a534b24b6494319a43857cf9b" Oct 03 13:21:07 crc kubenswrapper[4578]: I1003 13:21:07.491586 4578 scope.go:117] "RemoveContainer" containerID="1dc12dff3e12a4279cd27e0b3cfcb19a393615a985c7380d38cf793c967c9e1a" Oct 03 13:21:07 crc kubenswrapper[4578]: I1003 13:21:07.512802 4578 scope.go:117] "RemoveContainer" containerID="03725b4989bcd78d56fe85a2a5a02fb423b38d935130808036ccd0327ba40e2e" Oct 03 13:21:07 crc kubenswrapper[4578]: I1003 13:21:07.548013 4578 scope.go:117] "RemoveContainer" containerID="2fa6108292cf4712139d664537bc28e717b7be6deb73829c486d76891ae137f1" Oct 03 13:21:07 crc kubenswrapper[4578]: I1003 13:21:07.587505 4578 scope.go:117] "RemoveContainer" containerID="2903e87c0cb090ddcda043a9f1ec22e7ed66c8e3fbf355cc766c6bc247368eae" Oct 03 13:21:07 crc kubenswrapper[4578]: I1003 13:21:07.608195 4578 scope.go:117] "RemoveContainer" containerID="88a63d3cbf44911329d647457f6985e80beb070f3eea72c841db8dca447a6e10" Oct 03 13:21:12 crc kubenswrapper[4578]: I1003 13:21:12.910565 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:21:14 crc kubenswrapper[4578]: I1003 13:21:14.155770 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"a24ab26554e9c0b3f820f990bf2faaf8463fb8d2c4accf1896125b33a30b99b0"} Oct 03 13:21:18 crc kubenswrapper[4578]: I1003 13:21:18.054088 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-6l7bj"] Oct 03 13:21:18 crc kubenswrapper[4578]: I1003 13:21:18.064064 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-6l7bj"] Oct 03 13:21:18 crc kubenswrapper[4578]: I1003 13:21:18.920366 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8d4460d-feb5-4d93-95e4-260f5e06ee93" path="/var/lib/kubelet/pods/c8d4460d-feb5-4d93-95e4-260f5e06ee93/volumes" Oct 03 13:21:23 crc kubenswrapper[4578]: I1003 13:21:23.039395 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-xhpcq"] Oct 03 13:21:23 crc kubenswrapper[4578]: I1003 13:21:23.050126 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-xhpcq"] Oct 03 13:21:24 crc kubenswrapper[4578]: I1003 13:21:24.920613 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35460854-45fb-4080-aad3-3638a4283374" path="/var/lib/kubelet/pods/35460854-45fb-4080-aad3-3638a4283374/volumes" Oct 03 13:21:27 crc kubenswrapper[4578]: I1003 13:21:27.027326 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-mz57g"] Oct 03 13:21:27 crc kubenswrapper[4578]: I1003 13:21:27.037866 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-mz57g"] Oct 03 13:21:28 crc kubenswrapper[4578]: I1003 13:21:28.027660 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-f1e5-account-create-cvm99"] Oct 03 13:21:28 crc kubenswrapper[4578]: I1003 13:21:28.037823 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-f1e5-account-create-cvm99"] Oct 03 13:21:28 crc kubenswrapper[4578]: I1003 13:21:28.921298 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc" path="/var/lib/kubelet/pods/9ee3fb9c-6fa4-4217-b262-904ebc1ef6fc/volumes" Oct 03 13:21:28 crc kubenswrapper[4578]: I1003 13:21:28.923264 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad779b56-3577-401f-aba9-fc3b77d43200" path="/var/lib/kubelet/pods/ad779b56-3577-401f-aba9-fc3b77d43200/volumes" Oct 03 13:21:33 crc kubenswrapper[4578]: I1003 13:21:33.035109 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-950f-account-create-br2hk"] Oct 03 13:21:33 crc kubenswrapper[4578]: I1003 13:21:33.044430 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-950f-account-create-br2hk"] Oct 03 13:21:34 crc kubenswrapper[4578]: I1003 13:21:34.343847 4578 generic.go:334] "Generic (PLEG): container finished" podID="0095f720-7bf0-40fe-8f0c-e70fa5fc65d4" containerID="f86d2b3f43591a95ad325feb387f015bf4621162b162c0e2724af06df0ca54e1" exitCode=0 Oct 03 13:21:34 crc kubenswrapper[4578]: I1003 13:21:34.343955 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" event={"ID":"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4","Type":"ContainerDied","Data":"f86d2b3f43591a95ad325feb387f015bf4621162b162c0e2724af06df0ca54e1"} Oct 03 13:21:34 crc kubenswrapper[4578]: I1003 13:21:34.940209 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12a533e8-e9f7-49a6-ab49-f969cf8fd70b" path="/var/lib/kubelet/pods/12a533e8-e9f7-49a6-ab49-f969cf8fd70b/volumes" Oct 03 13:21:35 crc kubenswrapper[4578]: I1003 13:21:35.778909 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" Oct 03 13:21:35 crc kubenswrapper[4578]: I1003 13:21:35.920799 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-inventory\") pod \"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4\" (UID: \"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4\") " Oct 03 13:21:35 crc kubenswrapper[4578]: I1003 13:21:35.920961 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-bootstrap-combined-ca-bundle\") pod \"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4\" (UID: \"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4\") " Oct 03 13:21:35 crc kubenswrapper[4578]: I1003 13:21:35.921071 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w58gn\" (UniqueName: \"kubernetes.io/projected/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-kube-api-access-w58gn\") pod \"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4\" (UID: \"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4\") " Oct 03 13:21:35 crc kubenswrapper[4578]: I1003 13:21:35.921134 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-ssh-key\") pod \"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4\" (UID: \"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4\") " Oct 03 13:21:35 crc kubenswrapper[4578]: I1003 13:21:35.945564 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "0095f720-7bf0-40fe-8f0c-e70fa5fc65d4" (UID: "0095f720-7bf0-40fe-8f0c-e70fa5fc65d4"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:21:35 crc kubenswrapper[4578]: I1003 13:21:35.951021 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-kube-api-access-w58gn" (OuterVolumeSpecName: "kube-api-access-w58gn") pod "0095f720-7bf0-40fe-8f0c-e70fa5fc65d4" (UID: "0095f720-7bf0-40fe-8f0c-e70fa5fc65d4"). InnerVolumeSpecName "kube-api-access-w58gn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:21:35 crc kubenswrapper[4578]: I1003 13:21:35.960693 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-inventory" (OuterVolumeSpecName: "inventory") pod "0095f720-7bf0-40fe-8f0c-e70fa5fc65d4" (UID: "0095f720-7bf0-40fe-8f0c-e70fa5fc65d4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:21:35 crc kubenswrapper[4578]: I1003 13:21:35.966851 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0095f720-7bf0-40fe-8f0c-e70fa5fc65d4" (UID: "0095f720-7bf0-40fe-8f0c-e70fa5fc65d4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.023335 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w58gn\" (UniqueName: \"kubernetes.io/projected/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-kube-api-access-w58gn\") on node \"crc\" DevicePath \"\"" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.023357 4578 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.023366 4578 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.023375 4578 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0095f720-7bf0-40fe-8f0c-e70fa5fc65d4-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.363343 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" event={"ID":"0095f720-7bf0-40fe-8f0c-e70fa5fc65d4","Type":"ContainerDied","Data":"b27a7f7137642c2e3e6f5717c103e202f0ab9cc254b2f882ef523cfe928fb7f5"} Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.363416 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b27a7f7137642c2e3e6f5717c103e202f0ab9cc254b2f882ef523cfe928fb7f5" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.363475 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.461166 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg"] Oct 03 13:21:36 crc kubenswrapper[4578]: E1003 13:21:36.462045 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31df414b-9e4d-4a89-be61-17c99820ce42" containerName="registry-server" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.462062 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="31df414b-9e4d-4a89-be61-17c99820ce42" containerName="registry-server" Oct 03 13:21:36 crc kubenswrapper[4578]: E1003 13:21:36.462076 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62126af2-4cc0-4ccd-bb45-2520abe0f53d" containerName="extract-content" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.462084 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="62126af2-4cc0-4ccd-bb45-2520abe0f53d" containerName="extract-content" Oct 03 13:21:36 crc kubenswrapper[4578]: E1003 13:21:36.462103 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0095f720-7bf0-40fe-8f0c-e70fa5fc65d4" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.462112 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="0095f720-7bf0-40fe-8f0c-e70fa5fc65d4" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 03 13:21:36 crc kubenswrapper[4578]: E1003 13:21:36.462133 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62126af2-4cc0-4ccd-bb45-2520abe0f53d" containerName="extract-utilities" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.462140 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="62126af2-4cc0-4ccd-bb45-2520abe0f53d" containerName="extract-utilities" Oct 03 13:21:36 crc kubenswrapper[4578]: E1003 13:21:36.462153 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31df414b-9e4d-4a89-be61-17c99820ce42" containerName="extract-utilities" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.462160 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="31df414b-9e4d-4a89-be61-17c99820ce42" containerName="extract-utilities" Oct 03 13:21:36 crc kubenswrapper[4578]: E1003 13:21:36.462173 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62126af2-4cc0-4ccd-bb45-2520abe0f53d" containerName="registry-server" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.462180 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="62126af2-4cc0-4ccd-bb45-2520abe0f53d" containerName="registry-server" Oct 03 13:21:36 crc kubenswrapper[4578]: E1003 13:21:36.462205 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31df414b-9e4d-4a89-be61-17c99820ce42" containerName="extract-content" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.462212 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="31df414b-9e4d-4a89-be61-17c99820ce42" containerName="extract-content" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.462623 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="62126af2-4cc0-4ccd-bb45-2520abe0f53d" containerName="registry-server" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.462663 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="31df414b-9e4d-4a89-be61-17c99820ce42" containerName="registry-server" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.462673 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="0095f720-7bf0-40fe-8f0c-e70fa5fc65d4" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.463412 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.466149 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.466419 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-442mn" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.466588 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.466913 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.497373 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg"] Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.531865 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79b337ea-fd3a-49a6-9f82-67e1a86b29b0-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg\" (UID: \"79b337ea-fd3a-49a6-9f82-67e1a86b29b0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.531924 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzbs5\" (UniqueName: \"kubernetes.io/projected/79b337ea-fd3a-49a6-9f82-67e1a86b29b0-kube-api-access-fzbs5\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg\" (UID: \"79b337ea-fd3a-49a6-9f82-67e1a86b29b0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.531994 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79b337ea-fd3a-49a6-9f82-67e1a86b29b0-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg\" (UID: \"79b337ea-fd3a-49a6-9f82-67e1a86b29b0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.633963 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79b337ea-fd3a-49a6-9f82-67e1a86b29b0-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg\" (UID: \"79b337ea-fd3a-49a6-9f82-67e1a86b29b0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.634034 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzbs5\" (UniqueName: \"kubernetes.io/projected/79b337ea-fd3a-49a6-9f82-67e1a86b29b0-kube-api-access-fzbs5\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg\" (UID: \"79b337ea-fd3a-49a6-9f82-67e1a86b29b0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.634100 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79b337ea-fd3a-49a6-9f82-67e1a86b29b0-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg\" (UID: \"79b337ea-fd3a-49a6-9f82-67e1a86b29b0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.638303 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79b337ea-fd3a-49a6-9f82-67e1a86b29b0-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg\" (UID: \"79b337ea-fd3a-49a6-9f82-67e1a86b29b0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.639138 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79b337ea-fd3a-49a6-9f82-67e1a86b29b0-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg\" (UID: \"79b337ea-fd3a-49a6-9f82-67e1a86b29b0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.652448 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzbs5\" (UniqueName: \"kubernetes.io/projected/79b337ea-fd3a-49a6-9f82-67e1a86b29b0-kube-api-access-fzbs5\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg\" (UID: \"79b337ea-fd3a-49a6-9f82-67e1a86b29b0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg" Oct 03 13:21:36 crc kubenswrapper[4578]: I1003 13:21:36.795486 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg" Oct 03 13:21:37 crc kubenswrapper[4578]: I1003 13:21:37.310558 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg"] Oct 03 13:21:37 crc kubenswrapper[4578]: I1003 13:21:37.324816 4578 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 13:21:37 crc kubenswrapper[4578]: I1003 13:21:37.374190 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg" event={"ID":"79b337ea-fd3a-49a6-9f82-67e1a86b29b0","Type":"ContainerStarted","Data":"1a965d9ef20188e2ec908a712a315018b1b2c24d8cf58305362b020be62ff227"} Oct 03 13:21:38 crc kubenswrapper[4578]: I1003 13:21:38.025776 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-538c-account-create-kk4z8"] Oct 03 13:21:38 crc kubenswrapper[4578]: I1003 13:21:38.034472 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-538c-account-create-kk4z8"] Oct 03 13:21:38 crc kubenswrapper[4578]: I1003 13:21:38.382946 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg" event={"ID":"79b337ea-fd3a-49a6-9f82-67e1a86b29b0","Type":"ContainerStarted","Data":"c9883f8fa64aed65b319cb39542fa064b78d702ee8e18d9399f66cb1063d204c"} Oct 03 13:21:38 crc kubenswrapper[4578]: I1003 13:21:38.401887 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg" podStartSLOduration=2.271185469 podStartE2EDuration="2.401872849s" podCreationTimestamp="2025-10-03 13:21:36 +0000 UTC" firstStartedPulling="2025-10-03 13:21:37.324595907 +0000 UTC m=+1833.123068091" lastFinishedPulling="2025-10-03 13:21:37.455283287 +0000 UTC m=+1833.253755471" observedRunningTime="2025-10-03 13:21:38.399699732 +0000 UTC m=+1834.198171916" watchObservedRunningTime="2025-10-03 13:21:38.401872849 +0000 UTC m=+1834.200345033" Oct 03 13:21:38 crc kubenswrapper[4578]: I1003 13:21:38.921693 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="618c7600-498d-48ba-afd8-5b9d81638236" path="/var/lib/kubelet/pods/618c7600-498d-48ba-afd8-5b9d81638236/volumes" Oct 03 13:21:55 crc kubenswrapper[4578]: I1003 13:21:55.031606 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-wsqj9"] Oct 03 13:21:55 crc kubenswrapper[4578]: I1003 13:21:55.039171 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-vzh7p"] Oct 03 13:21:55 crc kubenswrapper[4578]: I1003 13:21:55.048996 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-twh27"] Oct 03 13:21:55 crc kubenswrapper[4578]: I1003 13:21:55.057068 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-wsqj9"] Oct 03 13:21:55 crc kubenswrapper[4578]: I1003 13:21:55.066160 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-vzh7p"] Oct 03 13:21:55 crc kubenswrapper[4578]: I1003 13:21:55.074911 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-twh27"] Oct 03 13:21:56 crc kubenswrapper[4578]: I1003 13:21:56.921246 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1953665a-f20b-4620-8061-bc3e4a33aacd" path="/var/lib/kubelet/pods/1953665a-f20b-4620-8061-bc3e4a33aacd/volumes" Oct 03 13:21:56 crc kubenswrapper[4578]: I1003 13:21:56.925021 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8999650e-96de-4898-9589-156a996042ed" path="/var/lib/kubelet/pods/8999650e-96de-4898-9589-156a996042ed/volumes" Oct 03 13:21:56 crc kubenswrapper[4578]: I1003 13:21:56.927092 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="beb0dafe-f425-4969-b586-29ca36b936f0" path="/var/lib/kubelet/pods/beb0dafe-f425-4969-b586-29ca36b936f0/volumes" Oct 03 13:22:02 crc kubenswrapper[4578]: I1003 13:22:02.036125 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-x6xhq"] Oct 03 13:22:02 crc kubenswrapper[4578]: I1003 13:22:02.043684 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-x6xhq"] Oct 03 13:22:02 crc kubenswrapper[4578]: I1003 13:22:02.921316 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7542b422-3e5f-479c-ae79-d3209197d088" path="/var/lib/kubelet/pods/7542b422-3e5f-479c-ae79-d3209197d088/volumes" Oct 03 13:22:05 crc kubenswrapper[4578]: I1003 13:22:05.044143 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-ad4f-account-create-bp75p"] Oct 03 13:22:05 crc kubenswrapper[4578]: I1003 13:22:05.056834 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-f504-account-create-vrmwv"] Oct 03 13:22:05 crc kubenswrapper[4578]: I1003 13:22:05.064206 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-9znjc"] Oct 03 13:22:05 crc kubenswrapper[4578]: I1003 13:22:05.071522 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-f504-account-create-vrmwv"] Oct 03 13:22:05 crc kubenswrapper[4578]: I1003 13:22:05.079158 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-ad4f-account-create-bp75p"] Oct 03 13:22:05 crc kubenswrapper[4578]: I1003 13:22:05.087197 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-2bce-account-create-zg8r5"] Oct 03 13:22:05 crc kubenswrapper[4578]: I1003 13:22:05.095769 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-9znjc"] Oct 03 13:22:05 crc kubenswrapper[4578]: I1003 13:22:05.102808 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-2bce-account-create-zg8r5"] Oct 03 13:22:06 crc kubenswrapper[4578]: I1003 13:22:06.919810 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3beaf630-6678-4fd2-8dc7-a6e50f51066c" path="/var/lib/kubelet/pods/3beaf630-6678-4fd2-8dc7-a6e50f51066c/volumes" Oct 03 13:22:06 crc kubenswrapper[4578]: I1003 13:22:06.922027 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ca2b580-f2b5-4f61-9aa8-63b114198bb0" path="/var/lib/kubelet/pods/6ca2b580-f2b5-4f61-9aa8-63b114198bb0/volumes" Oct 03 13:22:06 crc kubenswrapper[4578]: I1003 13:22:06.924305 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2537234-31d7-4970-a722-3e41b628008d" path="/var/lib/kubelet/pods/b2537234-31d7-4970-a722-3e41b628008d/volumes" Oct 03 13:22:06 crc kubenswrapper[4578]: I1003 13:22:06.925025 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6bc24ff-faa7-41fa-a16c-ddd9533e39f4" path="/var/lib/kubelet/pods/b6bc24ff-faa7-41fa-a16c-ddd9533e39f4/volumes" Oct 03 13:22:07 crc kubenswrapper[4578]: I1003 13:22:07.688574 4578 scope.go:117] "RemoveContainer" containerID="50e29480e63e2b503ccb4c85c9df0b0ac393219bfa25c4c5027a0e465c8efabd" Oct 03 13:22:07 crc kubenswrapper[4578]: I1003 13:22:07.723990 4578 scope.go:117] "RemoveContainer" containerID="7aa6fd6921703d6f49aedc22a01890766fc26f88404e28afecee61be9dca4e38" Oct 03 13:22:07 crc kubenswrapper[4578]: I1003 13:22:07.772226 4578 scope.go:117] "RemoveContainer" containerID="cff3994ec6b012ecee7843e150362954874e37adb08d33b5ce4811c1968777f0" Oct 03 13:22:07 crc kubenswrapper[4578]: I1003 13:22:07.841927 4578 scope.go:117] "RemoveContainer" containerID="97ba9510bcb1157d8687ff99064ab2bdcd89cf9a4d79983e75d04ef236835412" Oct 03 13:22:07 crc kubenswrapper[4578]: I1003 13:22:07.866050 4578 scope.go:117] "RemoveContainer" containerID="5e1761875604e3e19eafa1974e4b52f4c326151f1a6c0a086240eecae5b56378" Oct 03 13:22:07 crc kubenswrapper[4578]: I1003 13:22:07.903048 4578 scope.go:117] "RemoveContainer" containerID="5f0c1811e453f528295066dc5922a14afc246f86e36a0320719f0eac013a3cc8" Oct 03 13:22:07 crc kubenswrapper[4578]: I1003 13:22:07.941048 4578 scope.go:117] "RemoveContainer" containerID="86369178d83c39ee274194d59242ab148400430763c8e37c53b0b3a25eef58a8" Oct 03 13:22:07 crc kubenswrapper[4578]: I1003 13:22:07.959149 4578 scope.go:117] "RemoveContainer" containerID="42be6a1ce882316a4c4f29463fafa8e22e0b267c6d13f8a2c788e0532bf6e4ad" Oct 03 13:22:07 crc kubenswrapper[4578]: I1003 13:22:07.975862 4578 scope.go:117] "RemoveContainer" containerID="f46007ee12c93bed7b9fae98890c7330192ad2321fe90b5b2bf9dc8b84bb1357" Oct 03 13:22:07 crc kubenswrapper[4578]: I1003 13:22:07.994739 4578 scope.go:117] "RemoveContainer" containerID="7a8b2e6b02891c43dc3f8afd40babd3a6d66964e50f2d9ff378202f835c0debe" Oct 03 13:22:08 crc kubenswrapper[4578]: I1003 13:22:08.012460 4578 scope.go:117] "RemoveContainer" containerID="e1cbf44efb40d92f0202b86942cdf00f47b5fbd2338ecf5dcd010d207d07a59f" Oct 03 13:22:08 crc kubenswrapper[4578]: I1003 13:22:08.031158 4578 scope.go:117] "RemoveContainer" containerID="5dc6396f7f18ad1c0394a563d4a36cda87300d00170c0f6c07959c20c6032a6d" Oct 03 13:22:08 crc kubenswrapper[4578]: I1003 13:22:08.049809 4578 scope.go:117] "RemoveContainer" containerID="03a2337d0819cd158802e9abcef1d736a407794a36e8ae2f6f6451d58eae1aeb" Oct 03 13:22:08 crc kubenswrapper[4578]: I1003 13:22:08.074134 4578 scope.go:117] "RemoveContainer" containerID="4b057177e19710c3252908b4c28cc2e673d957c473c87d7dc0e2c3c5305d9647" Oct 03 13:22:08 crc kubenswrapper[4578]: I1003 13:22:08.102613 4578 scope.go:117] "RemoveContainer" containerID="a4c1d9d3955644651e82ef57819ea32c343727a4ac98c390148a5c47d256e96e" Oct 03 13:22:08 crc kubenswrapper[4578]: I1003 13:22:08.129472 4578 scope.go:117] "RemoveContainer" containerID="7e7ceab89b4db745e976243499d67a364246038e2af2277eaa10066965917e54" Oct 03 13:22:40 crc kubenswrapper[4578]: I1003 13:22:40.040807 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-85psh"] Oct 03 13:22:40 crc kubenswrapper[4578]: I1003 13:22:40.048587 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-85psh"] Oct 03 13:22:40 crc kubenswrapper[4578]: I1003 13:22:40.922272 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924" path="/var/lib/kubelet/pods/0ed8ad5f-dc9d-4b7e-9433-47f7a0c84924/volumes" Oct 03 13:22:50 crc kubenswrapper[4578]: I1003 13:22:50.033428 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-gv9w6"] Oct 03 13:22:50 crc kubenswrapper[4578]: I1003 13:22:50.041310 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-gv9w6"] Oct 03 13:22:50 crc kubenswrapper[4578]: I1003 13:22:50.920009 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb6557a6-8665-4a81-afec-4ccaef43b801" path="/var/lib/kubelet/pods/eb6557a6-8665-4a81-afec-4ccaef43b801/volumes" Oct 03 13:22:55 crc kubenswrapper[4578]: I1003 13:22:55.023770 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-j9scz"] Oct 03 13:22:55 crc kubenswrapper[4578]: I1003 13:22:55.033460 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-j9scz"] Oct 03 13:22:56 crc kubenswrapper[4578]: I1003 13:22:56.921603 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eacf42c3-8e6f-4304-b17b-de32359e8d76" path="/var/lib/kubelet/pods/eacf42c3-8e6f-4304-b17b-de32359e8d76/volumes" Oct 03 13:23:00 crc kubenswrapper[4578]: I1003 13:23:00.029125 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-jrdpv"] Oct 03 13:23:00 crc kubenswrapper[4578]: I1003 13:23:00.039575 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-jrdpv"] Oct 03 13:23:00 crc kubenswrapper[4578]: I1003 13:23:00.921129 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9856a93-43ca-4f3d-96a5-50c85086ae91" path="/var/lib/kubelet/pods/a9856a93-43ca-4f3d-96a5-50c85086ae91/volumes" Oct 03 13:23:08 crc kubenswrapper[4578]: I1003 13:23:08.392180 4578 scope.go:117] "RemoveContainer" containerID="5fa1c518576e9d2ffc9dda2b0ef5adaf2cf39618bda67a1eee0fb45f6bd0f0f4" Oct 03 13:23:08 crc kubenswrapper[4578]: I1003 13:23:08.421900 4578 scope.go:117] "RemoveContainer" containerID="4c31f2c7bec5ec73b4641d8ab5761d2f40e161d8db396a1f5ffe89700025a8c9" Oct 03 13:23:08 crc kubenswrapper[4578]: I1003 13:23:08.470093 4578 scope.go:117] "RemoveContainer" containerID="0abf83735c6600eb12b2aad4c07c5128d9e4fed23e86e7fec2aadf575ac91bcb" Oct 03 13:23:08 crc kubenswrapper[4578]: I1003 13:23:08.520836 4578 scope.go:117] "RemoveContainer" containerID="d3deb53df311f3445d8b678731a0915eeb05d092fbb84ef502995999cad84baf" Oct 03 13:23:21 crc kubenswrapper[4578]: I1003 13:23:21.040325 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-q6ppj"] Oct 03 13:23:21 crc kubenswrapper[4578]: I1003 13:23:21.048292 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-q6ppj"] Oct 03 13:23:21 crc kubenswrapper[4578]: I1003 13:23:21.253411 4578 generic.go:334] "Generic (PLEG): container finished" podID="79b337ea-fd3a-49a6-9f82-67e1a86b29b0" containerID="c9883f8fa64aed65b319cb39542fa064b78d702ee8e18d9399f66cb1063d204c" exitCode=0 Oct 03 13:23:21 crc kubenswrapper[4578]: I1003 13:23:21.253459 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg" event={"ID":"79b337ea-fd3a-49a6-9f82-67e1a86b29b0","Type":"ContainerDied","Data":"c9883f8fa64aed65b319cb39542fa064b78d702ee8e18d9399f66cb1063d204c"} Oct 03 13:23:22 crc kubenswrapper[4578]: I1003 13:23:22.636435 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg" Oct 03 13:23:22 crc kubenswrapper[4578]: I1003 13:23:22.762018 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzbs5\" (UniqueName: \"kubernetes.io/projected/79b337ea-fd3a-49a6-9f82-67e1a86b29b0-kube-api-access-fzbs5\") pod \"79b337ea-fd3a-49a6-9f82-67e1a86b29b0\" (UID: \"79b337ea-fd3a-49a6-9f82-67e1a86b29b0\") " Oct 03 13:23:22 crc kubenswrapper[4578]: I1003 13:23:22.762143 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79b337ea-fd3a-49a6-9f82-67e1a86b29b0-inventory\") pod \"79b337ea-fd3a-49a6-9f82-67e1a86b29b0\" (UID: \"79b337ea-fd3a-49a6-9f82-67e1a86b29b0\") " Oct 03 13:23:22 crc kubenswrapper[4578]: I1003 13:23:22.762203 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79b337ea-fd3a-49a6-9f82-67e1a86b29b0-ssh-key\") pod \"79b337ea-fd3a-49a6-9f82-67e1a86b29b0\" (UID: \"79b337ea-fd3a-49a6-9f82-67e1a86b29b0\") " Oct 03 13:23:22 crc kubenswrapper[4578]: I1003 13:23:22.774842 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79b337ea-fd3a-49a6-9f82-67e1a86b29b0-kube-api-access-fzbs5" (OuterVolumeSpecName: "kube-api-access-fzbs5") pod "79b337ea-fd3a-49a6-9f82-67e1a86b29b0" (UID: "79b337ea-fd3a-49a6-9f82-67e1a86b29b0"). InnerVolumeSpecName "kube-api-access-fzbs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:23:22 crc kubenswrapper[4578]: I1003 13:23:22.791069 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79b337ea-fd3a-49a6-9f82-67e1a86b29b0-inventory" (OuterVolumeSpecName: "inventory") pod "79b337ea-fd3a-49a6-9f82-67e1a86b29b0" (UID: "79b337ea-fd3a-49a6-9f82-67e1a86b29b0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:23:22 crc kubenswrapper[4578]: I1003 13:23:22.793113 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79b337ea-fd3a-49a6-9f82-67e1a86b29b0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "79b337ea-fd3a-49a6-9f82-67e1a86b29b0" (UID: "79b337ea-fd3a-49a6-9f82-67e1a86b29b0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:23:22 crc kubenswrapper[4578]: I1003 13:23:22.863871 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzbs5\" (UniqueName: \"kubernetes.io/projected/79b337ea-fd3a-49a6-9f82-67e1a86b29b0-kube-api-access-fzbs5\") on node \"crc\" DevicePath \"\"" Oct 03 13:23:22 crc kubenswrapper[4578]: I1003 13:23:22.863904 4578 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79b337ea-fd3a-49a6-9f82-67e1a86b29b0-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 13:23:22 crc kubenswrapper[4578]: I1003 13:23:22.863916 4578 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79b337ea-fd3a-49a6-9f82-67e1a86b29b0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:23:22 crc kubenswrapper[4578]: I1003 13:23:22.919391 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b310034-f02f-48ac-a47a-4f41f15379a5" path="/var/lib/kubelet/pods/6b310034-f02f-48ac-a47a-4f41f15379a5/volumes" Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.276799 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg" event={"ID":"79b337ea-fd3a-49a6-9f82-67e1a86b29b0","Type":"ContainerDied","Data":"1a965d9ef20188e2ec908a712a315018b1b2c24d8cf58305362b020be62ff227"} Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.276853 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a965d9ef20188e2ec908a712a315018b1b2c24d8cf58305362b020be62ff227" Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.276936 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg" Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.348779 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv"] Oct 03 13:23:23 crc kubenswrapper[4578]: E1003 13:23:23.349157 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79b337ea-fd3a-49a6-9f82-67e1a86b29b0" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.349175 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="79b337ea-fd3a-49a6-9f82-67e1a86b29b0" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.349371 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="79b337ea-fd3a-49a6-9f82-67e1a86b29b0" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.350041 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv" Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.352309 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-442mn" Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.352513 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.352931 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.353704 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.369146 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv"] Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.477399 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6glj\" (UniqueName: \"kubernetes.io/projected/70bd4887-f262-43e1-bd62-2a4f1727c67b-kube-api-access-g6glj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv\" (UID: \"70bd4887-f262-43e1-bd62-2a4f1727c67b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv" Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.477478 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70bd4887-f262-43e1-bd62-2a4f1727c67b-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv\" (UID: \"70bd4887-f262-43e1-bd62-2a4f1727c67b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv" Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.477538 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70bd4887-f262-43e1-bd62-2a4f1727c67b-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv\" (UID: \"70bd4887-f262-43e1-bd62-2a4f1727c67b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv" Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.579799 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6glj\" (UniqueName: \"kubernetes.io/projected/70bd4887-f262-43e1-bd62-2a4f1727c67b-kube-api-access-g6glj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv\" (UID: \"70bd4887-f262-43e1-bd62-2a4f1727c67b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv" Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.579926 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70bd4887-f262-43e1-bd62-2a4f1727c67b-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv\" (UID: \"70bd4887-f262-43e1-bd62-2a4f1727c67b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv" Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.580014 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70bd4887-f262-43e1-bd62-2a4f1727c67b-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv\" (UID: \"70bd4887-f262-43e1-bd62-2a4f1727c67b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv" Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.585566 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70bd4887-f262-43e1-bd62-2a4f1727c67b-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv\" (UID: \"70bd4887-f262-43e1-bd62-2a4f1727c67b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv" Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.585596 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70bd4887-f262-43e1-bd62-2a4f1727c67b-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv\" (UID: \"70bd4887-f262-43e1-bd62-2a4f1727c67b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv" Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.603329 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6glj\" (UniqueName: \"kubernetes.io/projected/70bd4887-f262-43e1-bd62-2a4f1727c67b-kube-api-access-g6glj\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv\" (UID: \"70bd4887-f262-43e1-bd62-2a4f1727c67b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv" Oct 03 13:23:23 crc kubenswrapper[4578]: I1003 13:23:23.707446 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv" Oct 03 13:23:24 crc kubenswrapper[4578]: I1003 13:23:24.228669 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv"] Oct 03 13:23:24 crc kubenswrapper[4578]: I1003 13:23:24.286036 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv" event={"ID":"70bd4887-f262-43e1-bd62-2a4f1727c67b","Type":"ContainerStarted","Data":"2b83c91af893c61a0a1aea6a05dc94a8111d7da0f433d2158405a123f6a5d6b3"} Oct 03 13:23:25 crc kubenswrapper[4578]: I1003 13:23:25.295481 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv" event={"ID":"70bd4887-f262-43e1-bd62-2a4f1727c67b","Type":"ContainerStarted","Data":"395e419a2c2a8753739946cf7b3db3367f0c3355c4442594d18a330084e9c69f"} Oct 03 13:23:25 crc kubenswrapper[4578]: I1003 13:23:25.328268 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv" podStartSLOduration=2.18272863 podStartE2EDuration="2.32824594s" podCreationTimestamp="2025-10-03 13:23:23 +0000 UTC" firstStartedPulling="2025-10-03 13:23:24.231199271 +0000 UTC m=+1940.029671465" lastFinishedPulling="2025-10-03 13:23:24.376716591 +0000 UTC m=+1940.175188775" observedRunningTime="2025-10-03 13:23:25.313181233 +0000 UTC m=+1941.111653427" watchObservedRunningTime="2025-10-03 13:23:25.32824594 +0000 UTC m=+1941.126718114" Oct 03 13:23:35 crc kubenswrapper[4578]: I1003 13:23:35.091707 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:23:35 crc kubenswrapper[4578]: I1003 13:23:35.092228 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:23:59 crc kubenswrapper[4578]: I1003 13:23:59.040814 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-56phj"] Oct 03 13:23:59 crc kubenswrapper[4578]: I1003 13:23:59.049682 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-sdwwm"] Oct 03 13:23:59 crc kubenswrapper[4578]: I1003 13:23:59.058718 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-4t996"] Oct 03 13:23:59 crc kubenswrapper[4578]: I1003 13:23:59.065998 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-sdwwm"] Oct 03 13:23:59 crc kubenswrapper[4578]: I1003 13:23:59.072273 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-4t996"] Oct 03 13:23:59 crc kubenswrapper[4578]: I1003 13:23:59.078295 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-56phj"] Oct 03 13:24:00 crc kubenswrapper[4578]: I1003 13:24:00.921621 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5533c61e-c654-4d79-9431-88bc37cff6d8" path="/var/lib/kubelet/pods/5533c61e-c654-4d79-9431-88bc37cff6d8/volumes" Oct 03 13:24:00 crc kubenswrapper[4578]: I1003 13:24:00.924007 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f94584d-3f52-4dcc-9c36-7f1174df9753" path="/var/lib/kubelet/pods/9f94584d-3f52-4dcc-9c36-7f1174df9753/volumes" Oct 03 13:24:00 crc kubenswrapper[4578]: I1003 13:24:00.924864 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cafa819d-515a-4e7a-8f97-e97522c4b31c" path="/var/lib/kubelet/pods/cafa819d-515a-4e7a-8f97-e97522c4b31c/volumes" Oct 03 13:24:05 crc kubenswrapper[4578]: I1003 13:24:05.091931 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:24:05 crc kubenswrapper[4578]: I1003 13:24:05.092478 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:24:08 crc kubenswrapper[4578]: I1003 13:24:08.037966 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-9d21-account-create-47c2p"] Oct 03 13:24:08 crc kubenswrapper[4578]: I1003 13:24:08.050582 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-9d21-account-create-47c2p"] Oct 03 13:24:08 crc kubenswrapper[4578]: I1003 13:24:08.647187 4578 scope.go:117] "RemoveContainer" containerID="439f5b895a77cc3be3500cced64bdf342ce57baa36cd07c7c3e9563b9cacefce" Oct 03 13:24:08 crc kubenswrapper[4578]: I1003 13:24:08.670274 4578 scope.go:117] "RemoveContainer" containerID="6f40a44541545d416250934fd25814d456a7fe570a45d77ca7f4b22f8992f19d" Oct 03 13:24:08 crc kubenswrapper[4578]: I1003 13:24:08.714987 4578 scope.go:117] "RemoveContainer" containerID="4a7daca41c304b21cfd34068e7c4f88933b9c6c3148f24319a72504cd0d21be2" Oct 03 13:24:08 crc kubenswrapper[4578]: I1003 13:24:08.776358 4578 scope.go:117] "RemoveContainer" containerID="0381801bf87524ccc1641c4cf296156a2ee3996c771b45e4e60575214a418d27" Oct 03 13:24:08 crc kubenswrapper[4578]: I1003 13:24:08.924309 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9cd3675-bb86-4dbd-8da1-a25d0fe30298" path="/var/lib/kubelet/pods/c9cd3675-bb86-4dbd-8da1-a25d0fe30298/volumes" Oct 03 13:24:09 crc kubenswrapper[4578]: I1003 13:24:09.028079 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-b3d4-account-create-nxpl4"] Oct 03 13:24:09 crc kubenswrapper[4578]: I1003 13:24:09.036598 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-b3d4-account-create-nxpl4"] Oct 03 13:24:09 crc kubenswrapper[4578]: I1003 13:24:09.044886 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-5bf9-account-create-9vgfl"] Oct 03 13:24:09 crc kubenswrapper[4578]: I1003 13:24:09.052537 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-5bf9-account-create-9vgfl"] Oct 03 13:24:10 crc kubenswrapper[4578]: I1003 13:24:10.920101 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="166d9486-f455-4541-b53c-89b7203f6994" path="/var/lib/kubelet/pods/166d9486-f455-4541-b53c-89b7203f6994/volumes" Oct 03 13:24:10 crc kubenswrapper[4578]: I1003 13:24:10.920667 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4229058-7b31-4a35-8ddc-6db4cf3da084" path="/var/lib/kubelet/pods/a4229058-7b31-4a35-8ddc-6db4cf3da084/volumes" Oct 03 13:24:35 crc kubenswrapper[4578]: I1003 13:24:35.041020 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fmdjc"] Oct 03 13:24:35 crc kubenswrapper[4578]: I1003 13:24:35.049620 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fmdjc"] Oct 03 13:24:35 crc kubenswrapper[4578]: I1003 13:24:35.091710 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:24:35 crc kubenswrapper[4578]: I1003 13:24:35.091780 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:24:35 crc kubenswrapper[4578]: I1003 13:24:35.091822 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 13:24:35 crc kubenswrapper[4578]: I1003 13:24:35.092524 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a24ab26554e9c0b3f820f990bf2faaf8463fb8d2c4accf1896125b33a30b99b0"} pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:24:35 crc kubenswrapper[4578]: I1003 13:24:35.092594 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" containerID="cri-o://a24ab26554e9c0b3f820f990bf2faaf8463fb8d2c4accf1896125b33a30b99b0" gracePeriod=600 Oct 03 13:24:35 crc kubenswrapper[4578]: I1003 13:24:35.870232 4578 generic.go:334] "Generic (PLEG): container finished" podID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerID="a24ab26554e9c0b3f820f990bf2faaf8463fb8d2c4accf1896125b33a30b99b0" exitCode=0 Oct 03 13:24:35 crc kubenswrapper[4578]: I1003 13:24:35.870297 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerDied","Data":"a24ab26554e9c0b3f820f990bf2faaf8463fb8d2c4accf1896125b33a30b99b0"} Oct 03 13:24:35 crc kubenswrapper[4578]: I1003 13:24:35.870938 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7"} Oct 03 13:24:35 crc kubenswrapper[4578]: I1003 13:24:35.870974 4578 scope.go:117] "RemoveContainer" containerID="e47a788d23816e2a8c2a0564b1297514ac7eafbdf44c832bba03614ce1617b95" Oct 03 13:24:36 crc kubenswrapper[4578]: I1003 13:24:36.921139 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f6e7467-e546-4563-bbe5-30d181f18349" path="/var/lib/kubelet/pods/5f6e7467-e546-4563-bbe5-30d181f18349/volumes" Oct 03 13:24:43 crc kubenswrapper[4578]: I1003 13:24:43.945490 4578 generic.go:334] "Generic (PLEG): container finished" podID="70bd4887-f262-43e1-bd62-2a4f1727c67b" containerID="395e419a2c2a8753739946cf7b3db3367f0c3355c4442594d18a330084e9c69f" exitCode=0 Oct 03 13:24:43 crc kubenswrapper[4578]: I1003 13:24:43.945738 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv" event={"ID":"70bd4887-f262-43e1-bd62-2a4f1727c67b","Type":"ContainerDied","Data":"395e419a2c2a8753739946cf7b3db3367f0c3355c4442594d18a330084e9c69f"} Oct 03 13:24:45 crc kubenswrapper[4578]: I1003 13:24:45.351853 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv" Oct 03 13:24:45 crc kubenswrapper[4578]: I1003 13:24:45.530297 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70bd4887-f262-43e1-bd62-2a4f1727c67b-inventory\") pod \"70bd4887-f262-43e1-bd62-2a4f1727c67b\" (UID: \"70bd4887-f262-43e1-bd62-2a4f1727c67b\") " Oct 03 13:24:45 crc kubenswrapper[4578]: I1003 13:24:45.530421 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70bd4887-f262-43e1-bd62-2a4f1727c67b-ssh-key\") pod \"70bd4887-f262-43e1-bd62-2a4f1727c67b\" (UID: \"70bd4887-f262-43e1-bd62-2a4f1727c67b\") " Oct 03 13:24:45 crc kubenswrapper[4578]: I1003 13:24:45.530443 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6glj\" (UniqueName: \"kubernetes.io/projected/70bd4887-f262-43e1-bd62-2a4f1727c67b-kube-api-access-g6glj\") pod \"70bd4887-f262-43e1-bd62-2a4f1727c67b\" (UID: \"70bd4887-f262-43e1-bd62-2a4f1727c67b\") " Oct 03 13:24:45 crc kubenswrapper[4578]: I1003 13:24:45.538989 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70bd4887-f262-43e1-bd62-2a4f1727c67b-kube-api-access-g6glj" (OuterVolumeSpecName: "kube-api-access-g6glj") pod "70bd4887-f262-43e1-bd62-2a4f1727c67b" (UID: "70bd4887-f262-43e1-bd62-2a4f1727c67b"). InnerVolumeSpecName "kube-api-access-g6glj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:24:45 crc kubenswrapper[4578]: I1003 13:24:45.560981 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70bd4887-f262-43e1-bd62-2a4f1727c67b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "70bd4887-f262-43e1-bd62-2a4f1727c67b" (UID: "70bd4887-f262-43e1-bd62-2a4f1727c67b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:24:45 crc kubenswrapper[4578]: I1003 13:24:45.569083 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70bd4887-f262-43e1-bd62-2a4f1727c67b-inventory" (OuterVolumeSpecName: "inventory") pod "70bd4887-f262-43e1-bd62-2a4f1727c67b" (UID: "70bd4887-f262-43e1-bd62-2a4f1727c67b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:24:45 crc kubenswrapper[4578]: I1003 13:24:45.633266 4578 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70bd4887-f262-43e1-bd62-2a4f1727c67b-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 13:24:45 crc kubenswrapper[4578]: I1003 13:24:45.633537 4578 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70bd4887-f262-43e1-bd62-2a4f1727c67b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:24:45 crc kubenswrapper[4578]: I1003 13:24:45.633652 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6glj\" (UniqueName: \"kubernetes.io/projected/70bd4887-f262-43e1-bd62-2a4f1727c67b-kube-api-access-g6glj\") on node \"crc\" DevicePath \"\"" Oct 03 13:24:45 crc kubenswrapper[4578]: I1003 13:24:45.961972 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv" event={"ID":"70bd4887-f262-43e1-bd62-2a4f1727c67b","Type":"ContainerDied","Data":"2b83c91af893c61a0a1aea6a05dc94a8111d7da0f433d2158405a123f6a5d6b3"} Oct 03 13:24:45 crc kubenswrapper[4578]: I1003 13:24:45.962311 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b83c91af893c61a0a1aea6a05dc94a8111d7da0f433d2158405a123f6a5d6b3" Oct 03 13:24:45 crc kubenswrapper[4578]: I1003 13:24:45.962408 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv" Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.042237 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9"] Oct 03 13:24:46 crc kubenswrapper[4578]: E1003 13:24:46.044706 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70bd4887-f262-43e1-bd62-2a4f1727c67b" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.044814 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="70bd4887-f262-43e1-bd62-2a4f1727c67b" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.045121 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="70bd4887-f262-43e1-bd62-2a4f1727c67b" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.045925 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9" Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.050123 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.050168 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.053859 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-442mn" Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.053971 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.056819 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9"] Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.146810 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a2213e2-d0c1-4481-898e-c1d9a52164d6-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9\" (UID: \"7a2213e2-d0c1-4481-898e-c1d9a52164d6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9" Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.146890 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a2213e2-d0c1-4481-898e-c1d9a52164d6-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9\" (UID: \"7a2213e2-d0c1-4481-898e-c1d9a52164d6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9" Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.147074 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvwzx\" (UniqueName: \"kubernetes.io/projected/7a2213e2-d0c1-4481-898e-c1d9a52164d6-kube-api-access-pvwzx\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9\" (UID: \"7a2213e2-d0c1-4481-898e-c1d9a52164d6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9" Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.249676 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvwzx\" (UniqueName: \"kubernetes.io/projected/7a2213e2-d0c1-4481-898e-c1d9a52164d6-kube-api-access-pvwzx\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9\" (UID: \"7a2213e2-d0c1-4481-898e-c1d9a52164d6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9" Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.249769 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a2213e2-d0c1-4481-898e-c1d9a52164d6-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9\" (UID: \"7a2213e2-d0c1-4481-898e-c1d9a52164d6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9" Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.249813 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a2213e2-d0c1-4481-898e-c1d9a52164d6-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9\" (UID: \"7a2213e2-d0c1-4481-898e-c1d9a52164d6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9" Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.256259 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a2213e2-d0c1-4481-898e-c1d9a52164d6-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9\" (UID: \"7a2213e2-d0c1-4481-898e-c1d9a52164d6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9" Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.257910 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a2213e2-d0c1-4481-898e-c1d9a52164d6-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9\" (UID: \"7a2213e2-d0c1-4481-898e-c1d9a52164d6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9" Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.272219 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvwzx\" (UniqueName: \"kubernetes.io/projected/7a2213e2-d0c1-4481-898e-c1d9a52164d6-kube-api-access-pvwzx\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9\" (UID: \"7a2213e2-d0c1-4481-898e-c1d9a52164d6\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9" Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.365193 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9" Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.921436 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9"] Oct 03 13:24:46 crc kubenswrapper[4578]: I1003 13:24:46.969969 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9" event={"ID":"7a2213e2-d0c1-4481-898e-c1d9a52164d6","Type":"ContainerStarted","Data":"479b7760da16e03188f50b5c57f304bbc2427c8cbdac623e3fa710433ea24a65"} Oct 03 13:24:47 crc kubenswrapper[4578]: I1003 13:24:47.979053 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9" event={"ID":"7a2213e2-d0c1-4481-898e-c1d9a52164d6","Type":"ContainerStarted","Data":"7f364d84cc9e2b2bba1be734a11b1ca3232c618455c990185c2e1919f27d2378"} Oct 03 13:24:48 crc kubenswrapper[4578]: I1003 13:24:48.001890 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9" podStartSLOduration=1.8647679209999999 podStartE2EDuration="2.001872579s" podCreationTimestamp="2025-10-03 13:24:46 +0000 UTC" firstStartedPulling="2025-10-03 13:24:46.933093698 +0000 UTC m=+2022.731565892" lastFinishedPulling="2025-10-03 13:24:47.070198366 +0000 UTC m=+2022.868670550" observedRunningTime="2025-10-03 13:24:48.000649631 +0000 UTC m=+2023.799121815" watchObservedRunningTime="2025-10-03 13:24:48.001872579 +0000 UTC m=+2023.800344763" Oct 03 13:24:53 crc kubenswrapper[4578]: I1003 13:24:53.035055 4578 generic.go:334] "Generic (PLEG): container finished" podID="7a2213e2-d0c1-4481-898e-c1d9a52164d6" containerID="7f364d84cc9e2b2bba1be734a11b1ca3232c618455c990185c2e1919f27d2378" exitCode=0 Oct 03 13:24:53 crc kubenswrapper[4578]: I1003 13:24:53.035087 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9" event={"ID":"7a2213e2-d0c1-4481-898e-c1d9a52164d6","Type":"ContainerDied","Data":"7f364d84cc9e2b2bba1be734a11b1ca3232c618455c990185c2e1919f27d2378"} Oct 03 13:24:54 crc kubenswrapper[4578]: I1003 13:24:54.459600 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9" Oct 03 13:24:54 crc kubenswrapper[4578]: I1003 13:24:54.603555 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a2213e2-d0c1-4481-898e-c1d9a52164d6-ssh-key\") pod \"7a2213e2-d0c1-4481-898e-c1d9a52164d6\" (UID: \"7a2213e2-d0c1-4481-898e-c1d9a52164d6\") " Oct 03 13:24:54 crc kubenswrapper[4578]: I1003 13:24:54.603758 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvwzx\" (UniqueName: \"kubernetes.io/projected/7a2213e2-d0c1-4481-898e-c1d9a52164d6-kube-api-access-pvwzx\") pod \"7a2213e2-d0c1-4481-898e-c1d9a52164d6\" (UID: \"7a2213e2-d0c1-4481-898e-c1d9a52164d6\") " Oct 03 13:24:54 crc kubenswrapper[4578]: I1003 13:24:54.603908 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a2213e2-d0c1-4481-898e-c1d9a52164d6-inventory\") pod \"7a2213e2-d0c1-4481-898e-c1d9a52164d6\" (UID: \"7a2213e2-d0c1-4481-898e-c1d9a52164d6\") " Oct 03 13:24:54 crc kubenswrapper[4578]: I1003 13:24:54.613547 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a2213e2-d0c1-4481-898e-c1d9a52164d6-kube-api-access-pvwzx" (OuterVolumeSpecName: "kube-api-access-pvwzx") pod "7a2213e2-d0c1-4481-898e-c1d9a52164d6" (UID: "7a2213e2-d0c1-4481-898e-c1d9a52164d6"). InnerVolumeSpecName "kube-api-access-pvwzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:24:54 crc kubenswrapper[4578]: I1003 13:24:54.636929 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a2213e2-d0c1-4481-898e-c1d9a52164d6-inventory" (OuterVolumeSpecName: "inventory") pod "7a2213e2-d0c1-4481-898e-c1d9a52164d6" (UID: "7a2213e2-d0c1-4481-898e-c1d9a52164d6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:24:54 crc kubenswrapper[4578]: I1003 13:24:54.637291 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a2213e2-d0c1-4481-898e-c1d9a52164d6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7a2213e2-d0c1-4481-898e-c1d9a52164d6" (UID: "7a2213e2-d0c1-4481-898e-c1d9a52164d6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:24:54 crc kubenswrapper[4578]: I1003 13:24:54.706232 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvwzx\" (UniqueName: \"kubernetes.io/projected/7a2213e2-d0c1-4481-898e-c1d9a52164d6-kube-api-access-pvwzx\") on node \"crc\" DevicePath \"\"" Oct 03 13:24:54 crc kubenswrapper[4578]: I1003 13:24:54.706266 4578 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7a2213e2-d0c1-4481-898e-c1d9a52164d6-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 13:24:54 crc kubenswrapper[4578]: I1003 13:24:54.706275 4578 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7a2213e2-d0c1-4481-898e-c1d9a52164d6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.051927 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9" event={"ID":"7a2213e2-d0c1-4481-898e-c1d9a52164d6","Type":"ContainerDied","Data":"479b7760da16e03188f50b5c57f304bbc2427c8cbdac623e3fa710433ea24a65"} Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.051965 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="479b7760da16e03188f50b5c57f304bbc2427c8cbdac623e3fa710433ea24a65" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.051986 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.132756 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm"] Oct 03 13:24:55 crc kubenswrapper[4578]: E1003 13:24:55.133186 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a2213e2-d0c1-4481-898e-c1d9a52164d6" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.133201 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a2213e2-d0c1-4481-898e-c1d9a52164d6" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.133399 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a2213e2-d0c1-4481-898e-c1d9a52164d6" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.134020 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.136264 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.137462 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.137690 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-442mn" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.137875 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.148172 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm"] Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.219871 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jnfmm\" (UID: \"ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.219968 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jnfmm\" (UID: \"ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.220031 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9hbj\" (UniqueName: \"kubernetes.io/projected/ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4-kube-api-access-p9hbj\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jnfmm\" (UID: \"ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.321499 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jnfmm\" (UID: \"ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.321605 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jnfmm\" (UID: \"ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.321677 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9hbj\" (UniqueName: \"kubernetes.io/projected/ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4-kube-api-access-p9hbj\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jnfmm\" (UID: \"ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.324964 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jnfmm\" (UID: \"ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.324964 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jnfmm\" (UID: \"ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.339254 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9hbj\" (UniqueName: \"kubernetes.io/projected/ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4-kube-api-access-p9hbj\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jnfmm\" (UID: \"ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.450237 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm" Oct 03 13:24:55 crc kubenswrapper[4578]: I1003 13:24:55.985592 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm"] Oct 03 13:24:56 crc kubenswrapper[4578]: I1003 13:24:56.063146 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm" event={"ID":"ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4","Type":"ContainerStarted","Data":"4dcb645e93f93382915fc70f05aac42f324373d3084a71f51f1ca357303e6405"} Oct 03 13:24:57 crc kubenswrapper[4578]: I1003 13:24:57.072005 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm" event={"ID":"ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4","Type":"ContainerStarted","Data":"dcf3acca5970b80bcf86264ed582f554cfcb5488d9f6ca69401b221e1cb30d13"} Oct 03 13:24:57 crc kubenswrapper[4578]: I1003 13:24:57.094162 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm" podStartSLOduration=1.944568885 podStartE2EDuration="2.094145031s" podCreationTimestamp="2025-10-03 13:24:55 +0000 UTC" firstStartedPulling="2025-10-03 13:24:55.994444209 +0000 UTC m=+2031.792916393" lastFinishedPulling="2025-10-03 13:24:56.144020355 +0000 UTC m=+2031.942492539" observedRunningTime="2025-10-03 13:24:57.087216256 +0000 UTC m=+2032.885688450" watchObservedRunningTime="2025-10-03 13:24:57.094145031 +0000 UTC m=+2032.892617215" Oct 03 13:25:01 crc kubenswrapper[4578]: I1003 13:25:01.041746 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-q4j79"] Oct 03 13:25:01 crc kubenswrapper[4578]: I1003 13:25:01.051005 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-q4j79"] Oct 03 13:25:02 crc kubenswrapper[4578]: I1003 13:25:02.028614 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-nvwxn"] Oct 03 13:25:02 crc kubenswrapper[4578]: I1003 13:25:02.035794 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-nvwxn"] Oct 03 13:25:02 crc kubenswrapper[4578]: I1003 13:25:02.919276 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e" path="/var/lib/kubelet/pods/1d1b3fe6-d8a6-450e-bbcf-5f910284ce2e/volumes" Oct 03 13:25:02 crc kubenswrapper[4578]: I1003 13:25:02.919858 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4304b326-aece-4557-904b-31f2fbab5c1f" path="/var/lib/kubelet/pods/4304b326-aece-4557-904b-31f2fbab5c1f/volumes" Oct 03 13:25:08 crc kubenswrapper[4578]: I1003 13:25:08.888447 4578 scope.go:117] "RemoveContainer" containerID="003a0c1928553820972cefea96c5d5af55961caad7bbb301c6b076d814788eec" Oct 03 13:25:08 crc kubenswrapper[4578]: I1003 13:25:08.925724 4578 scope.go:117] "RemoveContainer" containerID="242a9c8a497def7c5a681fa38a2aedd4beda4d7b738bbbbb43de23ec150a1fcb" Oct 03 13:25:08 crc kubenswrapper[4578]: I1003 13:25:08.980969 4578 scope.go:117] "RemoveContainer" containerID="2048be9c4225e7db96e0b1128fcfe04007983ee58ba570f83e79eebe69ed52b5" Oct 03 13:25:09 crc kubenswrapper[4578]: I1003 13:25:09.054493 4578 scope.go:117] "RemoveContainer" containerID="33b8b84c54d509a4a1b32d0b2f39a2180de18bdc686c351fb9eb8af094028901" Oct 03 13:25:09 crc kubenswrapper[4578]: I1003 13:25:09.090010 4578 scope.go:117] "RemoveContainer" containerID="359723667c8d4494fbfb2c97e0d75ca723efa35033aff65f5acddc2efdfab4f6" Oct 03 13:25:09 crc kubenswrapper[4578]: I1003 13:25:09.136815 4578 scope.go:117] "RemoveContainer" containerID="eba716b6694fb5f7bc63e249972dd056be854b91df8c6c802660a17d59055da2" Oct 03 13:25:34 crc kubenswrapper[4578]: I1003 13:25:34.397622 4578 generic.go:334] "Generic (PLEG): container finished" podID="ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4" containerID="dcf3acca5970b80bcf86264ed582f554cfcb5488d9f6ca69401b221e1cb30d13" exitCode=0 Oct 03 13:25:34 crc kubenswrapper[4578]: I1003 13:25:34.397758 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm" event={"ID":"ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4","Type":"ContainerDied","Data":"dcf3acca5970b80bcf86264ed582f554cfcb5488d9f6ca69401b221e1cb30d13"} Oct 03 13:25:35 crc kubenswrapper[4578]: I1003 13:25:35.714874 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wd8wr"] Oct 03 13:25:35 crc kubenswrapper[4578]: I1003 13:25:35.717220 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wd8wr" Oct 03 13:25:35 crc kubenswrapper[4578]: I1003 13:25:35.725211 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wd8wr"] Oct 03 13:25:35 crc kubenswrapper[4578]: I1003 13:25:35.796065 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm" Oct 03 13:25:35 crc kubenswrapper[4578]: I1003 13:25:35.853198 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnqdb\" (UniqueName: \"kubernetes.io/projected/a4c71279-daec-4485-8b17-fd954732b34c-kube-api-access-nnqdb\") pod \"community-operators-wd8wr\" (UID: \"a4c71279-daec-4485-8b17-fd954732b34c\") " pod="openshift-marketplace/community-operators-wd8wr" Oct 03 13:25:35 crc kubenswrapper[4578]: I1003 13:25:35.855331 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4c71279-daec-4485-8b17-fd954732b34c-catalog-content\") pod \"community-operators-wd8wr\" (UID: \"a4c71279-daec-4485-8b17-fd954732b34c\") " pod="openshift-marketplace/community-operators-wd8wr" Oct 03 13:25:35 crc kubenswrapper[4578]: I1003 13:25:35.855529 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4c71279-daec-4485-8b17-fd954732b34c-utilities\") pod \"community-operators-wd8wr\" (UID: \"a4c71279-daec-4485-8b17-fd954732b34c\") " pod="openshift-marketplace/community-operators-wd8wr" Oct 03 13:25:35 crc kubenswrapper[4578]: I1003 13:25:35.960340 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4-ssh-key\") pod \"ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4\" (UID: \"ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4\") " Oct 03 13:25:35 crc kubenswrapper[4578]: I1003 13:25:35.960413 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4-inventory\") pod \"ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4\" (UID: \"ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4\") " Oct 03 13:25:35 crc kubenswrapper[4578]: I1003 13:25:35.960527 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9hbj\" (UniqueName: \"kubernetes.io/projected/ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4-kube-api-access-p9hbj\") pod \"ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4\" (UID: \"ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4\") " Oct 03 13:25:35 crc kubenswrapper[4578]: I1003 13:25:35.960805 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnqdb\" (UniqueName: \"kubernetes.io/projected/a4c71279-daec-4485-8b17-fd954732b34c-kube-api-access-nnqdb\") pod \"community-operators-wd8wr\" (UID: \"a4c71279-daec-4485-8b17-fd954732b34c\") " pod="openshift-marketplace/community-operators-wd8wr" Oct 03 13:25:35 crc kubenswrapper[4578]: I1003 13:25:35.960846 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4c71279-daec-4485-8b17-fd954732b34c-catalog-content\") pod \"community-operators-wd8wr\" (UID: \"a4c71279-daec-4485-8b17-fd954732b34c\") " pod="openshift-marketplace/community-operators-wd8wr" Oct 03 13:25:35 crc kubenswrapper[4578]: I1003 13:25:35.960924 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4c71279-daec-4485-8b17-fd954732b34c-utilities\") pod \"community-operators-wd8wr\" (UID: \"a4c71279-daec-4485-8b17-fd954732b34c\") " pod="openshift-marketplace/community-operators-wd8wr" Oct 03 13:25:35 crc kubenswrapper[4578]: I1003 13:25:35.961448 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4c71279-daec-4485-8b17-fd954732b34c-utilities\") pod \"community-operators-wd8wr\" (UID: \"a4c71279-daec-4485-8b17-fd954732b34c\") " pod="openshift-marketplace/community-operators-wd8wr" Oct 03 13:25:35 crc kubenswrapper[4578]: I1003 13:25:35.961733 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4c71279-daec-4485-8b17-fd954732b34c-catalog-content\") pod \"community-operators-wd8wr\" (UID: \"a4c71279-daec-4485-8b17-fd954732b34c\") " pod="openshift-marketplace/community-operators-wd8wr" Oct 03 13:25:35 crc kubenswrapper[4578]: I1003 13:25:35.968744 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4-kube-api-access-p9hbj" (OuterVolumeSpecName: "kube-api-access-p9hbj") pod "ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4" (UID: "ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4"). InnerVolumeSpecName "kube-api-access-p9hbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:25:35 crc kubenswrapper[4578]: I1003 13:25:35.984481 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnqdb\" (UniqueName: \"kubernetes.io/projected/a4c71279-daec-4485-8b17-fd954732b34c-kube-api-access-nnqdb\") pod \"community-operators-wd8wr\" (UID: \"a4c71279-daec-4485-8b17-fd954732b34c\") " pod="openshift-marketplace/community-operators-wd8wr" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.000711 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4" (UID: "ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.037044 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4-inventory" (OuterVolumeSpecName: "inventory") pod "ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4" (UID: "ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.062882 4578 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.062918 4578 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.062930 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9hbj\" (UniqueName: \"kubernetes.io/projected/ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4-kube-api-access-p9hbj\") on node \"crc\" DevicePath \"\"" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.113202 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wd8wr" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.413542 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm" event={"ID":"ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4","Type":"ContainerDied","Data":"4dcb645e93f93382915fc70f05aac42f324373d3084a71f51f1ca357303e6405"} Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.413880 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4dcb645e93f93382915fc70f05aac42f324373d3084a71f51f1ca357303e6405" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.413945 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jnfmm" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.513566 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg"] Oct 03 13:25:36 crc kubenswrapper[4578]: E1003 13:25:36.513981 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.513996 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.514195 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.514901 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.519121 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-442mn" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.519368 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.519485 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.521353 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.535750 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg"] Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.656992 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wd8wr"] Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.686281 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg\" (UID: \"6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.686350 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg\" (UID: \"6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.686405 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxcs8\" (UniqueName: \"kubernetes.io/projected/6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e-kube-api-access-hxcs8\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg\" (UID: \"6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.788123 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxcs8\" (UniqueName: \"kubernetes.io/projected/6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e-kube-api-access-hxcs8\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg\" (UID: \"6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.788313 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg\" (UID: \"6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.788337 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg\" (UID: \"6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.793483 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg\" (UID: \"6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.793528 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg\" (UID: \"6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.806475 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxcs8\" (UniqueName: \"kubernetes.io/projected/6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e-kube-api-access-hxcs8\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg\" (UID: \"6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg" Oct 03 13:25:36 crc kubenswrapper[4578]: I1003 13:25:36.835851 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg" Oct 03 13:25:37 crc kubenswrapper[4578]: I1003 13:25:37.343743 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg"] Oct 03 13:25:37 crc kubenswrapper[4578]: W1003 13:25:37.346167 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b63f0ce_edc6_4405_8ec4_7d1b9c08dd1e.slice/crio-38dbf92b4c556549abb0fc342712170846ff8f94b7bd9021a23c11daa96431b4 WatchSource:0}: Error finding container 38dbf92b4c556549abb0fc342712170846ff8f94b7bd9021a23c11daa96431b4: Status 404 returned error can't find the container with id 38dbf92b4c556549abb0fc342712170846ff8f94b7bd9021a23c11daa96431b4 Oct 03 13:25:37 crc kubenswrapper[4578]: I1003 13:25:37.421635 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg" event={"ID":"6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e","Type":"ContainerStarted","Data":"38dbf92b4c556549abb0fc342712170846ff8f94b7bd9021a23c11daa96431b4"} Oct 03 13:25:37 crc kubenswrapper[4578]: I1003 13:25:37.423412 4578 generic.go:334] "Generic (PLEG): container finished" podID="a4c71279-daec-4485-8b17-fd954732b34c" containerID="f158c768366ef414b6cd0ef22a2bca6dd2d255a57a87510c8ae38c4f8b42b205" exitCode=0 Oct 03 13:25:37 crc kubenswrapper[4578]: I1003 13:25:37.423438 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd8wr" event={"ID":"a4c71279-daec-4485-8b17-fd954732b34c","Type":"ContainerDied","Data":"f158c768366ef414b6cd0ef22a2bca6dd2d255a57a87510c8ae38c4f8b42b205"} Oct 03 13:25:37 crc kubenswrapper[4578]: I1003 13:25:37.423457 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd8wr" event={"ID":"a4c71279-daec-4485-8b17-fd954732b34c","Type":"ContainerStarted","Data":"a8ba4d9c12dbd26f275857389366604bbe35d187465c3170cbd53212c00312bd"} Oct 03 13:25:38 crc kubenswrapper[4578]: I1003 13:25:38.437611 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd8wr" event={"ID":"a4c71279-daec-4485-8b17-fd954732b34c","Type":"ContainerStarted","Data":"80b1b6047585998c5337d2a2caa46071a845fef6ceb2acf77baa2e853dae5063"} Oct 03 13:25:38 crc kubenswrapper[4578]: I1003 13:25:38.440092 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg" event={"ID":"6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e","Type":"ContainerStarted","Data":"441fa300ed8e02970550f20b405b60a09f43b2974895f015047934e855b800ca"} Oct 03 13:25:38 crc kubenswrapper[4578]: I1003 13:25:38.481992 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg" podStartSLOduration=2.311640134 podStartE2EDuration="2.481968644s" podCreationTimestamp="2025-10-03 13:25:36 +0000 UTC" firstStartedPulling="2025-10-03 13:25:37.348378761 +0000 UTC m=+2073.146850945" lastFinishedPulling="2025-10-03 13:25:37.518707271 +0000 UTC m=+2073.317179455" observedRunningTime="2025-10-03 13:25:38.47312236 +0000 UTC m=+2074.271594554" watchObservedRunningTime="2025-10-03 13:25:38.481968644 +0000 UTC m=+2074.280440828" Oct 03 13:25:40 crc kubenswrapper[4578]: I1003 13:25:40.459853 4578 generic.go:334] "Generic (PLEG): container finished" podID="a4c71279-daec-4485-8b17-fd954732b34c" containerID="80b1b6047585998c5337d2a2caa46071a845fef6ceb2acf77baa2e853dae5063" exitCode=0 Oct 03 13:25:40 crc kubenswrapper[4578]: I1003 13:25:40.460042 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd8wr" event={"ID":"a4c71279-daec-4485-8b17-fd954732b34c","Type":"ContainerDied","Data":"80b1b6047585998c5337d2a2caa46071a845fef6ceb2acf77baa2e853dae5063"} Oct 03 13:25:41 crc kubenswrapper[4578]: I1003 13:25:41.472894 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd8wr" event={"ID":"a4c71279-daec-4485-8b17-fd954732b34c","Type":"ContainerStarted","Data":"8ba55713d5d6c8c96457c0884218d3ced96ba37ecc0d07c9edc9f10181020b46"} Oct 03 13:25:41 crc kubenswrapper[4578]: I1003 13:25:41.494419 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wd8wr" podStartSLOduration=2.7781402870000003 podStartE2EDuration="6.494399846s" podCreationTimestamp="2025-10-03 13:25:35 +0000 UTC" firstStartedPulling="2025-10-03 13:25:37.425051532 +0000 UTC m=+2073.223523716" lastFinishedPulling="2025-10-03 13:25:41.141311091 +0000 UTC m=+2076.939783275" observedRunningTime="2025-10-03 13:25:41.487687218 +0000 UTC m=+2077.286159412" watchObservedRunningTime="2025-10-03 13:25:41.494399846 +0000 UTC m=+2077.292872030" Oct 03 13:25:46 crc kubenswrapper[4578]: I1003 13:25:46.063062 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-mz6wt"] Oct 03 13:25:46 crc kubenswrapper[4578]: I1003 13:25:46.075867 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-mz6wt"] Oct 03 13:25:46 crc kubenswrapper[4578]: I1003 13:25:46.113714 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wd8wr" Oct 03 13:25:46 crc kubenswrapper[4578]: I1003 13:25:46.113760 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wd8wr" Oct 03 13:25:46 crc kubenswrapper[4578]: I1003 13:25:46.920714 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3e10cb8-af89-4561-a4ae-804943cf9440" path="/var/lib/kubelet/pods/d3e10cb8-af89-4561-a4ae-804943cf9440/volumes" Oct 03 13:25:47 crc kubenswrapper[4578]: I1003 13:25:47.165127 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-wd8wr" podUID="a4c71279-daec-4485-8b17-fd954732b34c" containerName="registry-server" probeResult="failure" output=< Oct 03 13:25:47 crc kubenswrapper[4578]: timeout: failed to connect service ":50051" within 1s Oct 03 13:25:47 crc kubenswrapper[4578]: > Oct 03 13:25:56 crc kubenswrapper[4578]: I1003 13:25:56.162535 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wd8wr" Oct 03 13:25:56 crc kubenswrapper[4578]: I1003 13:25:56.217515 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wd8wr" Oct 03 13:25:56 crc kubenswrapper[4578]: I1003 13:25:56.402700 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wd8wr"] Oct 03 13:25:57 crc kubenswrapper[4578]: I1003 13:25:57.610695 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wd8wr" podUID="a4c71279-daec-4485-8b17-fd954732b34c" containerName="registry-server" containerID="cri-o://8ba55713d5d6c8c96457c0884218d3ced96ba37ecc0d07c9edc9f10181020b46" gracePeriod=2 Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.049273 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wd8wr" Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.196223 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4c71279-daec-4485-8b17-fd954732b34c-utilities\") pod \"a4c71279-daec-4485-8b17-fd954732b34c\" (UID: \"a4c71279-daec-4485-8b17-fd954732b34c\") " Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.196382 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4c71279-daec-4485-8b17-fd954732b34c-catalog-content\") pod \"a4c71279-daec-4485-8b17-fd954732b34c\" (UID: \"a4c71279-daec-4485-8b17-fd954732b34c\") " Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.196468 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnqdb\" (UniqueName: \"kubernetes.io/projected/a4c71279-daec-4485-8b17-fd954732b34c-kube-api-access-nnqdb\") pod \"a4c71279-daec-4485-8b17-fd954732b34c\" (UID: \"a4c71279-daec-4485-8b17-fd954732b34c\") " Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.198259 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4c71279-daec-4485-8b17-fd954732b34c-utilities" (OuterVolumeSpecName: "utilities") pod "a4c71279-daec-4485-8b17-fd954732b34c" (UID: "a4c71279-daec-4485-8b17-fd954732b34c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.205086 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4c71279-daec-4485-8b17-fd954732b34c-kube-api-access-nnqdb" (OuterVolumeSpecName: "kube-api-access-nnqdb") pod "a4c71279-daec-4485-8b17-fd954732b34c" (UID: "a4c71279-daec-4485-8b17-fd954732b34c"). InnerVolumeSpecName "kube-api-access-nnqdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.248141 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4c71279-daec-4485-8b17-fd954732b34c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4c71279-daec-4485-8b17-fd954732b34c" (UID: "a4c71279-daec-4485-8b17-fd954732b34c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.298659 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4c71279-daec-4485-8b17-fd954732b34c-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.298692 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4c71279-daec-4485-8b17-fd954732b34c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.298704 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnqdb\" (UniqueName: \"kubernetes.io/projected/a4c71279-daec-4485-8b17-fd954732b34c-kube-api-access-nnqdb\") on node \"crc\" DevicePath \"\"" Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.623808 4578 generic.go:334] "Generic (PLEG): container finished" podID="a4c71279-daec-4485-8b17-fd954732b34c" containerID="8ba55713d5d6c8c96457c0884218d3ced96ba37ecc0d07c9edc9f10181020b46" exitCode=0 Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.623853 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd8wr" event={"ID":"a4c71279-daec-4485-8b17-fd954732b34c","Type":"ContainerDied","Data":"8ba55713d5d6c8c96457c0884218d3ced96ba37ecc0d07c9edc9f10181020b46"} Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.623885 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wd8wr" event={"ID":"a4c71279-daec-4485-8b17-fd954732b34c","Type":"ContainerDied","Data":"a8ba4d9c12dbd26f275857389366604bbe35d187465c3170cbd53212c00312bd"} Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.623892 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wd8wr" Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.623905 4578 scope.go:117] "RemoveContainer" containerID="8ba55713d5d6c8c96457c0884218d3ced96ba37ecc0d07c9edc9f10181020b46" Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.643470 4578 scope.go:117] "RemoveContainer" containerID="80b1b6047585998c5337d2a2caa46071a845fef6ceb2acf77baa2e853dae5063" Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.666856 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wd8wr"] Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.676261 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wd8wr"] Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.680560 4578 scope.go:117] "RemoveContainer" containerID="f158c768366ef414b6cd0ef22a2bca6dd2d255a57a87510c8ae38c4f8b42b205" Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.707932 4578 scope.go:117] "RemoveContainer" containerID="8ba55713d5d6c8c96457c0884218d3ced96ba37ecc0d07c9edc9f10181020b46" Oct 03 13:25:58 crc kubenswrapper[4578]: E1003 13:25:58.708392 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ba55713d5d6c8c96457c0884218d3ced96ba37ecc0d07c9edc9f10181020b46\": container with ID starting with 8ba55713d5d6c8c96457c0884218d3ced96ba37ecc0d07c9edc9f10181020b46 not found: ID does not exist" containerID="8ba55713d5d6c8c96457c0884218d3ced96ba37ecc0d07c9edc9f10181020b46" Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.708456 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ba55713d5d6c8c96457c0884218d3ced96ba37ecc0d07c9edc9f10181020b46"} err="failed to get container status \"8ba55713d5d6c8c96457c0884218d3ced96ba37ecc0d07c9edc9f10181020b46\": rpc error: code = NotFound desc = could not find container \"8ba55713d5d6c8c96457c0884218d3ced96ba37ecc0d07c9edc9f10181020b46\": container with ID starting with 8ba55713d5d6c8c96457c0884218d3ced96ba37ecc0d07c9edc9f10181020b46 not found: ID does not exist" Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.708511 4578 scope.go:117] "RemoveContainer" containerID="80b1b6047585998c5337d2a2caa46071a845fef6ceb2acf77baa2e853dae5063" Oct 03 13:25:58 crc kubenswrapper[4578]: E1003 13:25:58.708832 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80b1b6047585998c5337d2a2caa46071a845fef6ceb2acf77baa2e853dae5063\": container with ID starting with 80b1b6047585998c5337d2a2caa46071a845fef6ceb2acf77baa2e853dae5063 not found: ID does not exist" containerID="80b1b6047585998c5337d2a2caa46071a845fef6ceb2acf77baa2e853dae5063" Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.708889 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80b1b6047585998c5337d2a2caa46071a845fef6ceb2acf77baa2e853dae5063"} err="failed to get container status \"80b1b6047585998c5337d2a2caa46071a845fef6ceb2acf77baa2e853dae5063\": rpc error: code = NotFound desc = could not find container \"80b1b6047585998c5337d2a2caa46071a845fef6ceb2acf77baa2e853dae5063\": container with ID starting with 80b1b6047585998c5337d2a2caa46071a845fef6ceb2acf77baa2e853dae5063 not found: ID does not exist" Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.708910 4578 scope.go:117] "RemoveContainer" containerID="f158c768366ef414b6cd0ef22a2bca6dd2d255a57a87510c8ae38c4f8b42b205" Oct 03 13:25:58 crc kubenswrapper[4578]: E1003 13:25:58.709158 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f158c768366ef414b6cd0ef22a2bca6dd2d255a57a87510c8ae38c4f8b42b205\": container with ID starting with f158c768366ef414b6cd0ef22a2bca6dd2d255a57a87510c8ae38c4f8b42b205 not found: ID does not exist" containerID="f158c768366ef414b6cd0ef22a2bca6dd2d255a57a87510c8ae38c4f8b42b205" Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.709206 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f158c768366ef414b6cd0ef22a2bca6dd2d255a57a87510c8ae38c4f8b42b205"} err="failed to get container status \"f158c768366ef414b6cd0ef22a2bca6dd2d255a57a87510c8ae38c4f8b42b205\": rpc error: code = NotFound desc = could not find container \"f158c768366ef414b6cd0ef22a2bca6dd2d255a57a87510c8ae38c4f8b42b205\": container with ID starting with f158c768366ef414b6cd0ef22a2bca6dd2d255a57a87510c8ae38c4f8b42b205 not found: ID does not exist" Oct 03 13:25:58 crc kubenswrapper[4578]: I1003 13:25:58.921796 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4c71279-daec-4485-8b17-fd954732b34c" path="/var/lib/kubelet/pods/a4c71279-daec-4485-8b17-fd954732b34c/volumes" Oct 03 13:26:09 crc kubenswrapper[4578]: I1003 13:26:09.284250 4578 scope.go:117] "RemoveContainer" containerID="c4fd7d19b6ab5bea908d1b9c1dcde5509c71ba22e94c933a286e2563297c1ef8" Oct 03 13:26:32 crc kubenswrapper[4578]: I1003 13:26:32.869453 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c4mts"] Oct 03 13:26:32 crc kubenswrapper[4578]: E1003 13:26:32.870401 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4c71279-daec-4485-8b17-fd954732b34c" containerName="extract-utilities" Oct 03 13:26:32 crc kubenswrapper[4578]: I1003 13:26:32.870415 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4c71279-daec-4485-8b17-fd954732b34c" containerName="extract-utilities" Oct 03 13:26:32 crc kubenswrapper[4578]: E1003 13:26:32.870434 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4c71279-daec-4485-8b17-fd954732b34c" containerName="registry-server" Oct 03 13:26:32 crc kubenswrapper[4578]: I1003 13:26:32.870440 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4c71279-daec-4485-8b17-fd954732b34c" containerName="registry-server" Oct 03 13:26:32 crc kubenswrapper[4578]: E1003 13:26:32.870450 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4c71279-daec-4485-8b17-fd954732b34c" containerName="extract-content" Oct 03 13:26:32 crc kubenswrapper[4578]: I1003 13:26:32.870456 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4c71279-daec-4485-8b17-fd954732b34c" containerName="extract-content" Oct 03 13:26:32 crc kubenswrapper[4578]: I1003 13:26:32.870677 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4c71279-daec-4485-8b17-fd954732b34c" containerName="registry-server" Oct 03 13:26:32 crc kubenswrapper[4578]: I1003 13:26:32.872540 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4mts" Oct 03 13:26:32 crc kubenswrapper[4578]: I1003 13:26:32.887719 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c4mts"] Oct 03 13:26:32 crc kubenswrapper[4578]: I1003 13:26:32.923713 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dtrt\" (UniqueName: \"kubernetes.io/projected/393ef701-1ef3-4206-afdb-22973a0e6746-kube-api-access-7dtrt\") pod \"certified-operators-c4mts\" (UID: \"393ef701-1ef3-4206-afdb-22973a0e6746\") " pod="openshift-marketplace/certified-operators-c4mts" Oct 03 13:26:32 crc kubenswrapper[4578]: I1003 13:26:32.923971 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/393ef701-1ef3-4206-afdb-22973a0e6746-utilities\") pod \"certified-operators-c4mts\" (UID: \"393ef701-1ef3-4206-afdb-22973a0e6746\") " pod="openshift-marketplace/certified-operators-c4mts" Oct 03 13:26:32 crc kubenswrapper[4578]: I1003 13:26:32.924137 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/393ef701-1ef3-4206-afdb-22973a0e6746-catalog-content\") pod \"certified-operators-c4mts\" (UID: \"393ef701-1ef3-4206-afdb-22973a0e6746\") " pod="openshift-marketplace/certified-operators-c4mts" Oct 03 13:26:33 crc kubenswrapper[4578]: I1003 13:26:33.026186 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dtrt\" (UniqueName: \"kubernetes.io/projected/393ef701-1ef3-4206-afdb-22973a0e6746-kube-api-access-7dtrt\") pod \"certified-operators-c4mts\" (UID: \"393ef701-1ef3-4206-afdb-22973a0e6746\") " pod="openshift-marketplace/certified-operators-c4mts" Oct 03 13:26:33 crc kubenswrapper[4578]: I1003 13:26:33.026584 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/393ef701-1ef3-4206-afdb-22973a0e6746-utilities\") pod \"certified-operators-c4mts\" (UID: \"393ef701-1ef3-4206-afdb-22973a0e6746\") " pod="openshift-marketplace/certified-operators-c4mts" Oct 03 13:26:33 crc kubenswrapper[4578]: I1003 13:26:33.026921 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/393ef701-1ef3-4206-afdb-22973a0e6746-catalog-content\") pod \"certified-operators-c4mts\" (UID: \"393ef701-1ef3-4206-afdb-22973a0e6746\") " pod="openshift-marketplace/certified-operators-c4mts" Oct 03 13:26:33 crc kubenswrapper[4578]: I1003 13:26:33.027102 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/393ef701-1ef3-4206-afdb-22973a0e6746-utilities\") pod \"certified-operators-c4mts\" (UID: \"393ef701-1ef3-4206-afdb-22973a0e6746\") " pod="openshift-marketplace/certified-operators-c4mts" Oct 03 13:26:33 crc kubenswrapper[4578]: I1003 13:26:33.027428 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/393ef701-1ef3-4206-afdb-22973a0e6746-catalog-content\") pod \"certified-operators-c4mts\" (UID: \"393ef701-1ef3-4206-afdb-22973a0e6746\") " pod="openshift-marketplace/certified-operators-c4mts" Oct 03 13:26:33 crc kubenswrapper[4578]: I1003 13:26:33.051934 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dtrt\" (UniqueName: \"kubernetes.io/projected/393ef701-1ef3-4206-afdb-22973a0e6746-kube-api-access-7dtrt\") pod \"certified-operators-c4mts\" (UID: \"393ef701-1ef3-4206-afdb-22973a0e6746\") " pod="openshift-marketplace/certified-operators-c4mts" Oct 03 13:26:33 crc kubenswrapper[4578]: I1003 13:26:33.200610 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4mts" Oct 03 13:26:33 crc kubenswrapper[4578]: I1003 13:26:33.783212 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c4mts"] Oct 03 13:26:33 crc kubenswrapper[4578]: I1003 13:26:33.924525 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4mts" event={"ID":"393ef701-1ef3-4206-afdb-22973a0e6746","Type":"ContainerStarted","Data":"42e7b9ddaee087b8c50a54b0e4b954d0259c64cde63863b837a2c2747e14ecf3"} Oct 03 13:26:33 crc kubenswrapper[4578]: I1003 13:26:33.934763 4578 generic.go:334] "Generic (PLEG): container finished" podID="6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e" containerID="441fa300ed8e02970550f20b405b60a09f43b2974895f015047934e855b800ca" exitCode=2 Oct 03 13:26:33 crc kubenswrapper[4578]: I1003 13:26:33.934875 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg" event={"ID":"6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e","Type":"ContainerDied","Data":"441fa300ed8e02970550f20b405b60a09f43b2974895f015047934e855b800ca"} Oct 03 13:26:34 crc kubenswrapper[4578]: I1003 13:26:34.948043 4578 generic.go:334] "Generic (PLEG): container finished" podID="393ef701-1ef3-4206-afdb-22973a0e6746" containerID="2aeb860e86674a19c47659dcca6c658f93e43cacfea9cb637aaeea2b696961c4" exitCode=0 Oct 03 13:26:34 crc kubenswrapper[4578]: I1003 13:26:34.949699 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4mts" event={"ID":"393ef701-1ef3-4206-afdb-22973a0e6746","Type":"ContainerDied","Data":"2aeb860e86674a19c47659dcca6c658f93e43cacfea9cb637aaeea2b696961c4"} Oct 03 13:26:35 crc kubenswrapper[4578]: I1003 13:26:35.092230 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:26:35 crc kubenswrapper[4578]: I1003 13:26:35.092591 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:26:35 crc kubenswrapper[4578]: I1003 13:26:35.417700 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg" Oct 03 13:26:35 crc kubenswrapper[4578]: I1003 13:26:35.489652 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxcs8\" (UniqueName: \"kubernetes.io/projected/6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e-kube-api-access-hxcs8\") pod \"6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e\" (UID: \"6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e\") " Oct 03 13:26:35 crc kubenswrapper[4578]: I1003 13:26:35.489835 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e-inventory\") pod \"6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e\" (UID: \"6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e\") " Oct 03 13:26:35 crc kubenswrapper[4578]: I1003 13:26:35.489894 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e-ssh-key\") pod \"6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e\" (UID: \"6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e\") " Oct 03 13:26:35 crc kubenswrapper[4578]: I1003 13:26:35.495803 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e-kube-api-access-hxcs8" (OuterVolumeSpecName: "kube-api-access-hxcs8") pod "6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e" (UID: "6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e"). InnerVolumeSpecName "kube-api-access-hxcs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:26:35 crc kubenswrapper[4578]: I1003 13:26:35.520228 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e" (UID: "6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:26:35 crc kubenswrapper[4578]: I1003 13:26:35.521848 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e-inventory" (OuterVolumeSpecName: "inventory") pod "6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e" (UID: "6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:26:35 crc kubenswrapper[4578]: I1003 13:26:35.592233 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxcs8\" (UniqueName: \"kubernetes.io/projected/6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e-kube-api-access-hxcs8\") on node \"crc\" DevicePath \"\"" Oct 03 13:26:35 crc kubenswrapper[4578]: I1003 13:26:35.592277 4578 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 13:26:35 crc kubenswrapper[4578]: I1003 13:26:35.592289 4578 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:26:35 crc kubenswrapper[4578]: I1003 13:26:35.959146 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg" Oct 03 13:26:35 crc kubenswrapper[4578]: I1003 13:26:35.959483 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg" event={"ID":"6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e","Type":"ContainerDied","Data":"38dbf92b4c556549abb0fc342712170846ff8f94b7bd9021a23c11daa96431b4"} Oct 03 13:26:35 crc kubenswrapper[4578]: I1003 13:26:35.959524 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38dbf92b4c556549abb0fc342712170846ff8f94b7bd9021a23c11daa96431b4" Oct 03 13:26:36 crc kubenswrapper[4578]: I1003 13:26:36.980172 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4mts" event={"ID":"393ef701-1ef3-4206-afdb-22973a0e6746","Type":"ContainerStarted","Data":"08b438b9f5a12e26ebadc98b946fd0f3f9bbe8adf6de8fb03218e9e311fc2c84"} Oct 03 13:26:37 crc kubenswrapper[4578]: I1003 13:26:37.990375 4578 generic.go:334] "Generic (PLEG): container finished" podID="393ef701-1ef3-4206-afdb-22973a0e6746" containerID="08b438b9f5a12e26ebadc98b946fd0f3f9bbe8adf6de8fb03218e9e311fc2c84" exitCode=0 Oct 03 13:26:37 crc kubenswrapper[4578]: I1003 13:26:37.990439 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4mts" event={"ID":"393ef701-1ef3-4206-afdb-22973a0e6746","Type":"ContainerDied","Data":"08b438b9f5a12e26ebadc98b946fd0f3f9bbe8adf6de8fb03218e9e311fc2c84"} Oct 03 13:26:37 crc kubenswrapper[4578]: I1003 13:26:37.990667 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4mts" event={"ID":"393ef701-1ef3-4206-afdb-22973a0e6746","Type":"ContainerStarted","Data":"895681131bba961900f3ca3ecb8bf11c28ec97f662965127665cf3aa2459b2aa"} Oct 03 13:26:38 crc kubenswrapper[4578]: I1003 13:26:38.010951 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c4mts" podStartSLOduration=3.225148791 podStartE2EDuration="6.010930393s" podCreationTimestamp="2025-10-03 13:26:32 +0000 UTC" firstStartedPulling="2025-10-03 13:26:34.953809404 +0000 UTC m=+2130.752281588" lastFinishedPulling="2025-10-03 13:26:37.739591006 +0000 UTC m=+2133.538063190" observedRunningTime="2025-10-03 13:26:38.006795885 +0000 UTC m=+2133.805268079" watchObservedRunningTime="2025-10-03 13:26:38.010930393 +0000 UTC m=+2133.809402577" Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.031418 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt"] Oct 03 13:26:42 crc kubenswrapper[4578]: E1003 13:26:42.032735 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.032756 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.032967 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.033794 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt" Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.036588 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.037574 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.038613 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.040666 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-442mn" Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.045529 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt"] Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.107971 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tg4w\" (UniqueName: \"kubernetes.io/projected/53e7ff32-dd82-4be9-b309-7cd2b0130d32-kube-api-access-6tg4w\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7fctt\" (UID: \"53e7ff32-dd82-4be9-b309-7cd2b0130d32\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt" Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.108064 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53e7ff32-dd82-4be9-b309-7cd2b0130d32-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7fctt\" (UID: \"53e7ff32-dd82-4be9-b309-7cd2b0130d32\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt" Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.108136 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53e7ff32-dd82-4be9-b309-7cd2b0130d32-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7fctt\" (UID: \"53e7ff32-dd82-4be9-b309-7cd2b0130d32\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt" Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.209816 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tg4w\" (UniqueName: \"kubernetes.io/projected/53e7ff32-dd82-4be9-b309-7cd2b0130d32-kube-api-access-6tg4w\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7fctt\" (UID: \"53e7ff32-dd82-4be9-b309-7cd2b0130d32\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt" Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.209895 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53e7ff32-dd82-4be9-b309-7cd2b0130d32-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7fctt\" (UID: \"53e7ff32-dd82-4be9-b309-7cd2b0130d32\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt" Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.209965 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53e7ff32-dd82-4be9-b309-7cd2b0130d32-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7fctt\" (UID: \"53e7ff32-dd82-4be9-b309-7cd2b0130d32\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt" Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.216703 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53e7ff32-dd82-4be9-b309-7cd2b0130d32-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7fctt\" (UID: \"53e7ff32-dd82-4be9-b309-7cd2b0130d32\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt" Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.218030 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53e7ff32-dd82-4be9-b309-7cd2b0130d32-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7fctt\" (UID: \"53e7ff32-dd82-4be9-b309-7cd2b0130d32\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt" Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.231713 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tg4w\" (UniqueName: \"kubernetes.io/projected/53e7ff32-dd82-4be9-b309-7cd2b0130d32-kube-api-access-6tg4w\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-7fctt\" (UID: \"53e7ff32-dd82-4be9-b309-7cd2b0130d32\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt" Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.355864 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt" Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.916497 4578 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 13:26:42 crc kubenswrapper[4578]: I1003 13:26:42.923759 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt"] Oct 03 13:26:43 crc kubenswrapper[4578]: I1003 13:26:43.040976 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt" event={"ID":"53e7ff32-dd82-4be9-b309-7cd2b0130d32","Type":"ContainerStarted","Data":"2fa251747d07627c4bdab7f756c5b14aea0a967c5b0c428aee18b4a65c9ae90d"} Oct 03 13:26:43 crc kubenswrapper[4578]: I1003 13:26:43.200353 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c4mts" Oct 03 13:26:43 crc kubenswrapper[4578]: I1003 13:26:43.201706 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c4mts" Oct 03 13:26:43 crc kubenswrapper[4578]: I1003 13:26:43.265485 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c4mts" Oct 03 13:26:44 crc kubenswrapper[4578]: I1003 13:26:44.051883 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt" event={"ID":"53e7ff32-dd82-4be9-b309-7cd2b0130d32","Type":"ContainerStarted","Data":"2874ee2ea05c483e9e3838f0fc17da523f9775d25214b821b88d8e6d31f40817"} Oct 03 13:26:44 crc kubenswrapper[4578]: I1003 13:26:44.106039 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c4mts" Oct 03 13:26:44 crc kubenswrapper[4578]: I1003 13:26:44.130545 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt" podStartSLOduration=1.947735691 podStartE2EDuration="2.130526168s" podCreationTimestamp="2025-10-03 13:26:42 +0000 UTC" firstStartedPulling="2025-10-03 13:26:42.916312821 +0000 UTC m=+2138.714785005" lastFinishedPulling="2025-10-03 13:26:43.099103298 +0000 UTC m=+2138.897575482" observedRunningTime="2025-10-03 13:26:44.076176551 +0000 UTC m=+2139.874648745" watchObservedRunningTime="2025-10-03 13:26:44.130526168 +0000 UTC m=+2139.928998342" Oct 03 13:26:44 crc kubenswrapper[4578]: I1003 13:26:44.158302 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c4mts"] Oct 03 13:26:46 crc kubenswrapper[4578]: I1003 13:26:46.066268 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-c4mts" podUID="393ef701-1ef3-4206-afdb-22973a0e6746" containerName="registry-server" containerID="cri-o://895681131bba961900f3ca3ecb8bf11c28ec97f662965127665cf3aa2459b2aa" gracePeriod=2 Oct 03 13:26:46 crc kubenswrapper[4578]: I1003 13:26:46.524192 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4mts" Oct 03 13:26:46 crc kubenswrapper[4578]: I1003 13:26:46.603265 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/393ef701-1ef3-4206-afdb-22973a0e6746-utilities\") pod \"393ef701-1ef3-4206-afdb-22973a0e6746\" (UID: \"393ef701-1ef3-4206-afdb-22973a0e6746\") " Oct 03 13:26:46 crc kubenswrapper[4578]: I1003 13:26:46.603316 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dtrt\" (UniqueName: \"kubernetes.io/projected/393ef701-1ef3-4206-afdb-22973a0e6746-kube-api-access-7dtrt\") pod \"393ef701-1ef3-4206-afdb-22973a0e6746\" (UID: \"393ef701-1ef3-4206-afdb-22973a0e6746\") " Oct 03 13:26:46 crc kubenswrapper[4578]: I1003 13:26:46.603352 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/393ef701-1ef3-4206-afdb-22973a0e6746-catalog-content\") pod \"393ef701-1ef3-4206-afdb-22973a0e6746\" (UID: \"393ef701-1ef3-4206-afdb-22973a0e6746\") " Oct 03 13:26:46 crc kubenswrapper[4578]: I1003 13:26:46.610748 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/393ef701-1ef3-4206-afdb-22973a0e6746-utilities" (OuterVolumeSpecName: "utilities") pod "393ef701-1ef3-4206-afdb-22973a0e6746" (UID: "393ef701-1ef3-4206-afdb-22973a0e6746"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:26:46 crc kubenswrapper[4578]: I1003 13:26:46.632741 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/393ef701-1ef3-4206-afdb-22973a0e6746-kube-api-access-7dtrt" (OuterVolumeSpecName: "kube-api-access-7dtrt") pod "393ef701-1ef3-4206-afdb-22973a0e6746" (UID: "393ef701-1ef3-4206-afdb-22973a0e6746"). InnerVolumeSpecName "kube-api-access-7dtrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:26:46 crc kubenswrapper[4578]: I1003 13:26:46.645711 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/393ef701-1ef3-4206-afdb-22973a0e6746-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "393ef701-1ef3-4206-afdb-22973a0e6746" (UID: "393ef701-1ef3-4206-afdb-22973a0e6746"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:26:46 crc kubenswrapper[4578]: I1003 13:26:46.705227 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/393ef701-1ef3-4206-afdb-22973a0e6746-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:26:46 crc kubenswrapper[4578]: I1003 13:26:46.705269 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dtrt\" (UniqueName: \"kubernetes.io/projected/393ef701-1ef3-4206-afdb-22973a0e6746-kube-api-access-7dtrt\") on node \"crc\" DevicePath \"\"" Oct 03 13:26:46 crc kubenswrapper[4578]: I1003 13:26:46.705283 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/393ef701-1ef3-4206-afdb-22973a0e6746-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:26:47 crc kubenswrapper[4578]: I1003 13:26:47.074664 4578 generic.go:334] "Generic (PLEG): container finished" podID="393ef701-1ef3-4206-afdb-22973a0e6746" containerID="895681131bba961900f3ca3ecb8bf11c28ec97f662965127665cf3aa2459b2aa" exitCode=0 Oct 03 13:26:47 crc kubenswrapper[4578]: I1003 13:26:47.074719 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4mts" event={"ID":"393ef701-1ef3-4206-afdb-22973a0e6746","Type":"ContainerDied","Data":"895681131bba961900f3ca3ecb8bf11c28ec97f662965127665cf3aa2459b2aa"} Oct 03 13:26:47 crc kubenswrapper[4578]: I1003 13:26:47.074768 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4mts" event={"ID":"393ef701-1ef3-4206-afdb-22973a0e6746","Type":"ContainerDied","Data":"42e7b9ddaee087b8c50a54b0e4b954d0259c64cde63863b837a2c2747e14ecf3"} Oct 03 13:26:47 crc kubenswrapper[4578]: I1003 13:26:47.074791 4578 scope.go:117] "RemoveContainer" containerID="895681131bba961900f3ca3ecb8bf11c28ec97f662965127665cf3aa2459b2aa" Oct 03 13:26:47 crc kubenswrapper[4578]: I1003 13:26:47.075744 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4mts" Oct 03 13:26:47 crc kubenswrapper[4578]: I1003 13:26:47.098145 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c4mts"] Oct 03 13:26:47 crc kubenswrapper[4578]: I1003 13:26:47.101718 4578 scope.go:117] "RemoveContainer" containerID="08b438b9f5a12e26ebadc98b946fd0f3f9bbe8adf6de8fb03218e9e311fc2c84" Oct 03 13:26:47 crc kubenswrapper[4578]: I1003 13:26:47.106390 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-c4mts"] Oct 03 13:26:47 crc kubenswrapper[4578]: I1003 13:26:47.129699 4578 scope.go:117] "RemoveContainer" containerID="2aeb860e86674a19c47659dcca6c658f93e43cacfea9cb637aaeea2b696961c4" Oct 03 13:26:47 crc kubenswrapper[4578]: I1003 13:26:47.176393 4578 scope.go:117] "RemoveContainer" containerID="895681131bba961900f3ca3ecb8bf11c28ec97f662965127665cf3aa2459b2aa" Oct 03 13:26:47 crc kubenswrapper[4578]: E1003 13:26:47.176965 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"895681131bba961900f3ca3ecb8bf11c28ec97f662965127665cf3aa2459b2aa\": container with ID starting with 895681131bba961900f3ca3ecb8bf11c28ec97f662965127665cf3aa2459b2aa not found: ID does not exist" containerID="895681131bba961900f3ca3ecb8bf11c28ec97f662965127665cf3aa2459b2aa" Oct 03 13:26:47 crc kubenswrapper[4578]: I1003 13:26:47.177004 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"895681131bba961900f3ca3ecb8bf11c28ec97f662965127665cf3aa2459b2aa"} err="failed to get container status \"895681131bba961900f3ca3ecb8bf11c28ec97f662965127665cf3aa2459b2aa\": rpc error: code = NotFound desc = could not find container \"895681131bba961900f3ca3ecb8bf11c28ec97f662965127665cf3aa2459b2aa\": container with ID starting with 895681131bba961900f3ca3ecb8bf11c28ec97f662965127665cf3aa2459b2aa not found: ID does not exist" Oct 03 13:26:47 crc kubenswrapper[4578]: I1003 13:26:47.177029 4578 scope.go:117] "RemoveContainer" containerID="08b438b9f5a12e26ebadc98b946fd0f3f9bbe8adf6de8fb03218e9e311fc2c84" Oct 03 13:26:47 crc kubenswrapper[4578]: E1003 13:26:47.178050 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08b438b9f5a12e26ebadc98b946fd0f3f9bbe8adf6de8fb03218e9e311fc2c84\": container with ID starting with 08b438b9f5a12e26ebadc98b946fd0f3f9bbe8adf6de8fb03218e9e311fc2c84 not found: ID does not exist" containerID="08b438b9f5a12e26ebadc98b946fd0f3f9bbe8adf6de8fb03218e9e311fc2c84" Oct 03 13:26:47 crc kubenswrapper[4578]: I1003 13:26:47.178085 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08b438b9f5a12e26ebadc98b946fd0f3f9bbe8adf6de8fb03218e9e311fc2c84"} err="failed to get container status \"08b438b9f5a12e26ebadc98b946fd0f3f9bbe8adf6de8fb03218e9e311fc2c84\": rpc error: code = NotFound desc = could not find container \"08b438b9f5a12e26ebadc98b946fd0f3f9bbe8adf6de8fb03218e9e311fc2c84\": container with ID starting with 08b438b9f5a12e26ebadc98b946fd0f3f9bbe8adf6de8fb03218e9e311fc2c84 not found: ID does not exist" Oct 03 13:26:47 crc kubenswrapper[4578]: I1003 13:26:47.178108 4578 scope.go:117] "RemoveContainer" containerID="2aeb860e86674a19c47659dcca6c658f93e43cacfea9cb637aaeea2b696961c4" Oct 03 13:26:47 crc kubenswrapper[4578]: E1003 13:26:47.178813 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2aeb860e86674a19c47659dcca6c658f93e43cacfea9cb637aaeea2b696961c4\": container with ID starting with 2aeb860e86674a19c47659dcca6c658f93e43cacfea9cb637aaeea2b696961c4 not found: ID does not exist" containerID="2aeb860e86674a19c47659dcca6c658f93e43cacfea9cb637aaeea2b696961c4" Oct 03 13:26:47 crc kubenswrapper[4578]: I1003 13:26:47.178844 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2aeb860e86674a19c47659dcca6c658f93e43cacfea9cb637aaeea2b696961c4"} err="failed to get container status \"2aeb860e86674a19c47659dcca6c658f93e43cacfea9cb637aaeea2b696961c4\": rpc error: code = NotFound desc = could not find container \"2aeb860e86674a19c47659dcca6c658f93e43cacfea9cb637aaeea2b696961c4\": container with ID starting with 2aeb860e86674a19c47659dcca6c658f93e43cacfea9cb637aaeea2b696961c4 not found: ID does not exist" Oct 03 13:26:48 crc kubenswrapper[4578]: I1003 13:26:48.920504 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="393ef701-1ef3-4206-afdb-22973a0e6746" path="/var/lib/kubelet/pods/393ef701-1ef3-4206-afdb-22973a0e6746/volumes" Oct 03 13:27:05 crc kubenswrapper[4578]: I1003 13:27:05.091609 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:27:05 crc kubenswrapper[4578]: I1003 13:27:05.092165 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:27:34 crc kubenswrapper[4578]: I1003 13:27:34.447924 4578 generic.go:334] "Generic (PLEG): container finished" podID="53e7ff32-dd82-4be9-b309-7cd2b0130d32" containerID="2874ee2ea05c483e9e3838f0fc17da523f9775d25214b821b88d8e6d31f40817" exitCode=0 Oct 03 13:27:34 crc kubenswrapper[4578]: I1003 13:27:34.448014 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt" event={"ID":"53e7ff32-dd82-4be9-b309-7cd2b0130d32","Type":"ContainerDied","Data":"2874ee2ea05c483e9e3838f0fc17da523f9775d25214b821b88d8e6d31f40817"} Oct 03 13:27:35 crc kubenswrapper[4578]: I1003 13:27:35.091618 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:27:35 crc kubenswrapper[4578]: I1003 13:27:35.091714 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:27:35 crc kubenswrapper[4578]: I1003 13:27:35.091758 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 13:27:35 crc kubenswrapper[4578]: I1003 13:27:35.092459 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7"} pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:27:35 crc kubenswrapper[4578]: I1003 13:27:35.092512 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" containerID="cri-o://29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" gracePeriod=600 Oct 03 13:27:35 crc kubenswrapper[4578]: E1003 13:27:35.214966 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:27:35 crc kubenswrapper[4578]: I1003 13:27:35.458980 4578 generic.go:334] "Generic (PLEG): container finished" podID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" exitCode=0 Oct 03 13:27:35 crc kubenswrapper[4578]: I1003 13:27:35.459776 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerDied","Data":"29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7"} Oct 03 13:27:35 crc kubenswrapper[4578]: I1003 13:27:35.459845 4578 scope.go:117] "RemoveContainer" containerID="a24ab26554e9c0b3f820f990bf2faaf8463fb8d2c4accf1896125b33a30b99b0" Oct 03 13:27:35 crc kubenswrapper[4578]: I1003 13:27:35.460712 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:27:35 crc kubenswrapper[4578]: E1003 13:27:35.461173 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:27:35 crc kubenswrapper[4578]: I1003 13:27:35.895591 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.074173 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53e7ff32-dd82-4be9-b309-7cd2b0130d32-inventory\") pod \"53e7ff32-dd82-4be9-b309-7cd2b0130d32\" (UID: \"53e7ff32-dd82-4be9-b309-7cd2b0130d32\") " Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.074252 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53e7ff32-dd82-4be9-b309-7cd2b0130d32-ssh-key\") pod \"53e7ff32-dd82-4be9-b309-7cd2b0130d32\" (UID: \"53e7ff32-dd82-4be9-b309-7cd2b0130d32\") " Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.074283 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tg4w\" (UniqueName: \"kubernetes.io/projected/53e7ff32-dd82-4be9-b309-7cd2b0130d32-kube-api-access-6tg4w\") pod \"53e7ff32-dd82-4be9-b309-7cd2b0130d32\" (UID: \"53e7ff32-dd82-4be9-b309-7cd2b0130d32\") " Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.088831 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53e7ff32-dd82-4be9-b309-7cd2b0130d32-kube-api-access-6tg4w" (OuterVolumeSpecName: "kube-api-access-6tg4w") pod "53e7ff32-dd82-4be9-b309-7cd2b0130d32" (UID: "53e7ff32-dd82-4be9-b309-7cd2b0130d32"). InnerVolumeSpecName "kube-api-access-6tg4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.109252 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53e7ff32-dd82-4be9-b309-7cd2b0130d32-inventory" (OuterVolumeSpecName: "inventory") pod "53e7ff32-dd82-4be9-b309-7cd2b0130d32" (UID: "53e7ff32-dd82-4be9-b309-7cd2b0130d32"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.109899 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53e7ff32-dd82-4be9-b309-7cd2b0130d32-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "53e7ff32-dd82-4be9-b309-7cd2b0130d32" (UID: "53e7ff32-dd82-4be9-b309-7cd2b0130d32"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.176846 4578 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53e7ff32-dd82-4be9-b309-7cd2b0130d32-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.176879 4578 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53e7ff32-dd82-4be9-b309-7cd2b0130d32-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.176888 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tg4w\" (UniqueName: \"kubernetes.io/projected/53e7ff32-dd82-4be9-b309-7cd2b0130d32-kube-api-access-6tg4w\") on node \"crc\" DevicePath \"\"" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.472940 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt" event={"ID":"53e7ff32-dd82-4be9-b309-7cd2b0130d32","Type":"ContainerDied","Data":"2fa251747d07627c4bdab7f756c5b14aea0a967c5b0c428aee18b4a65c9ae90d"} Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.472988 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fa251747d07627c4bdab7f756c5b14aea0a967c5b0c428aee18b4a65c9ae90d" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.473052 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-7fctt" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.645799 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ggkhk"] Oct 03 13:27:36 crc kubenswrapper[4578]: E1003 13:27:36.646490 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="393ef701-1ef3-4206-afdb-22973a0e6746" containerName="registry-server" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.646512 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="393ef701-1ef3-4206-afdb-22973a0e6746" containerName="registry-server" Oct 03 13:27:36 crc kubenswrapper[4578]: E1003 13:27:36.646529 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="393ef701-1ef3-4206-afdb-22973a0e6746" containerName="extract-utilities" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.646536 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="393ef701-1ef3-4206-afdb-22973a0e6746" containerName="extract-utilities" Oct 03 13:27:36 crc kubenswrapper[4578]: E1003 13:27:36.646542 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="393ef701-1ef3-4206-afdb-22973a0e6746" containerName="extract-content" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.646548 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="393ef701-1ef3-4206-afdb-22973a0e6746" containerName="extract-content" Oct 03 13:27:36 crc kubenswrapper[4578]: E1003 13:27:36.646577 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53e7ff32-dd82-4be9-b309-7cd2b0130d32" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.646592 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="53e7ff32-dd82-4be9-b309-7cd2b0130d32" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.646798 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="53e7ff32-dd82-4be9-b309-7cd2b0130d32" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.646812 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="393ef701-1ef3-4206-afdb-22973a0e6746" containerName="registry-server" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.647406 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ggkhk" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.649768 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.650153 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.651735 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-442mn" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.652155 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.660491 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ggkhk"] Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.789950 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ac850868-8034-40e9-ac47-21455b1f8817-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ggkhk\" (UID: \"ac850868-8034-40e9-ac47-21455b1f8817\") " pod="openstack/ssh-known-hosts-edpm-deployment-ggkhk" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.790025 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zhhv\" (UniqueName: \"kubernetes.io/projected/ac850868-8034-40e9-ac47-21455b1f8817-kube-api-access-7zhhv\") pod \"ssh-known-hosts-edpm-deployment-ggkhk\" (UID: \"ac850868-8034-40e9-ac47-21455b1f8817\") " pod="openstack/ssh-known-hosts-edpm-deployment-ggkhk" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.790121 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ac850868-8034-40e9-ac47-21455b1f8817-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ggkhk\" (UID: \"ac850868-8034-40e9-ac47-21455b1f8817\") " pod="openstack/ssh-known-hosts-edpm-deployment-ggkhk" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.891691 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ac850868-8034-40e9-ac47-21455b1f8817-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ggkhk\" (UID: \"ac850868-8034-40e9-ac47-21455b1f8817\") " pod="openstack/ssh-known-hosts-edpm-deployment-ggkhk" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.891809 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zhhv\" (UniqueName: \"kubernetes.io/projected/ac850868-8034-40e9-ac47-21455b1f8817-kube-api-access-7zhhv\") pod \"ssh-known-hosts-edpm-deployment-ggkhk\" (UID: \"ac850868-8034-40e9-ac47-21455b1f8817\") " pod="openstack/ssh-known-hosts-edpm-deployment-ggkhk" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.891908 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ac850868-8034-40e9-ac47-21455b1f8817-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ggkhk\" (UID: \"ac850868-8034-40e9-ac47-21455b1f8817\") " pod="openstack/ssh-known-hosts-edpm-deployment-ggkhk" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.901418 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ac850868-8034-40e9-ac47-21455b1f8817-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ggkhk\" (UID: \"ac850868-8034-40e9-ac47-21455b1f8817\") " pod="openstack/ssh-known-hosts-edpm-deployment-ggkhk" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.901748 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ac850868-8034-40e9-ac47-21455b1f8817-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ggkhk\" (UID: \"ac850868-8034-40e9-ac47-21455b1f8817\") " pod="openstack/ssh-known-hosts-edpm-deployment-ggkhk" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.914450 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zhhv\" (UniqueName: \"kubernetes.io/projected/ac850868-8034-40e9-ac47-21455b1f8817-kube-api-access-7zhhv\") pod \"ssh-known-hosts-edpm-deployment-ggkhk\" (UID: \"ac850868-8034-40e9-ac47-21455b1f8817\") " pod="openstack/ssh-known-hosts-edpm-deployment-ggkhk" Oct 03 13:27:36 crc kubenswrapper[4578]: I1003 13:27:36.963219 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ggkhk" Oct 03 13:27:37 crc kubenswrapper[4578]: I1003 13:27:37.494468 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ggkhk"] Oct 03 13:27:37 crc kubenswrapper[4578]: W1003 13:27:37.496954 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac850868_8034_40e9_ac47_21455b1f8817.slice/crio-38d26929d0e0703f179da82216f81ca6b79400f065ed20bba12b7ff2518f041b WatchSource:0}: Error finding container 38d26929d0e0703f179da82216f81ca6b79400f065ed20bba12b7ff2518f041b: Status 404 returned error can't find the container with id 38d26929d0e0703f179da82216f81ca6b79400f065ed20bba12b7ff2518f041b Oct 03 13:27:38 crc kubenswrapper[4578]: I1003 13:27:38.494125 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ggkhk" event={"ID":"ac850868-8034-40e9-ac47-21455b1f8817","Type":"ContainerStarted","Data":"0da641d1c79799b6db757e54634bff86f9cbee86dbc66d0fa4b60613b4a7037d"} Oct 03 13:27:38 crc kubenswrapper[4578]: I1003 13:27:38.494511 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ggkhk" event={"ID":"ac850868-8034-40e9-ac47-21455b1f8817","Type":"ContainerStarted","Data":"38d26929d0e0703f179da82216f81ca6b79400f065ed20bba12b7ff2518f041b"} Oct 03 13:27:38 crc kubenswrapper[4578]: I1003 13:27:38.515107 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-ggkhk" podStartSLOduration=2.164008734 podStartE2EDuration="2.515087187s" podCreationTimestamp="2025-10-03 13:27:36 +0000 UTC" firstStartedPulling="2025-10-03 13:27:37.499136956 +0000 UTC m=+2193.297609140" lastFinishedPulling="2025-10-03 13:27:37.850215409 +0000 UTC m=+2193.648687593" observedRunningTime="2025-10-03 13:27:38.511071372 +0000 UTC m=+2194.309543556" watchObservedRunningTime="2025-10-03 13:27:38.515087187 +0000 UTC m=+2194.313559371" Oct 03 13:27:45 crc kubenswrapper[4578]: I1003 13:27:45.552239 4578 generic.go:334] "Generic (PLEG): container finished" podID="ac850868-8034-40e9-ac47-21455b1f8817" containerID="0da641d1c79799b6db757e54634bff86f9cbee86dbc66d0fa4b60613b4a7037d" exitCode=0 Oct 03 13:27:45 crc kubenswrapper[4578]: I1003 13:27:45.552320 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ggkhk" event={"ID":"ac850868-8034-40e9-ac47-21455b1f8817","Type":"ContainerDied","Data":"0da641d1c79799b6db757e54634bff86f9cbee86dbc66d0fa4b60613b4a7037d"} Oct 03 13:27:46 crc kubenswrapper[4578]: I1003 13:27:46.969608 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ggkhk" Oct 03 13:27:46 crc kubenswrapper[4578]: I1003 13:27:46.981136 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ac850868-8034-40e9-ac47-21455b1f8817-ssh-key-openstack-edpm-ipam\") pod \"ac850868-8034-40e9-ac47-21455b1f8817\" (UID: \"ac850868-8034-40e9-ac47-21455b1f8817\") " Oct 03 13:27:46 crc kubenswrapper[4578]: I1003 13:27:46.981219 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zhhv\" (UniqueName: \"kubernetes.io/projected/ac850868-8034-40e9-ac47-21455b1f8817-kube-api-access-7zhhv\") pod \"ac850868-8034-40e9-ac47-21455b1f8817\" (UID: \"ac850868-8034-40e9-ac47-21455b1f8817\") " Oct 03 13:27:46 crc kubenswrapper[4578]: I1003 13:27:46.981356 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ac850868-8034-40e9-ac47-21455b1f8817-inventory-0\") pod \"ac850868-8034-40e9-ac47-21455b1f8817\" (UID: \"ac850868-8034-40e9-ac47-21455b1f8817\") " Oct 03 13:27:46 crc kubenswrapper[4578]: I1003 13:27:46.992805 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac850868-8034-40e9-ac47-21455b1f8817-kube-api-access-7zhhv" (OuterVolumeSpecName: "kube-api-access-7zhhv") pod "ac850868-8034-40e9-ac47-21455b1f8817" (UID: "ac850868-8034-40e9-ac47-21455b1f8817"). InnerVolumeSpecName "kube-api-access-7zhhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.093337 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zhhv\" (UniqueName: \"kubernetes.io/projected/ac850868-8034-40e9-ac47-21455b1f8817-kube-api-access-7zhhv\") on node \"crc\" DevicePath \"\"" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.115953 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac850868-8034-40e9-ac47-21455b1f8817-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "ac850868-8034-40e9-ac47-21455b1f8817" (UID: "ac850868-8034-40e9-ac47-21455b1f8817"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.139822 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac850868-8034-40e9-ac47-21455b1f8817-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "ac850868-8034-40e9-ac47-21455b1f8817" (UID: "ac850868-8034-40e9-ac47-21455b1f8817"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.194827 4578 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ac850868-8034-40e9-ac47-21455b1f8817-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.194864 4578 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ac850868-8034-40e9-ac47-21455b1f8817-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.572482 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ggkhk" event={"ID":"ac850868-8034-40e9-ac47-21455b1f8817","Type":"ContainerDied","Data":"38d26929d0e0703f179da82216f81ca6b79400f065ed20bba12b7ff2518f041b"} Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.572531 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38d26929d0e0703f179da82216f81ca6b79400f065ed20bba12b7ff2518f041b" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.572537 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ggkhk" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.650707 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46"] Oct 03 13:27:47 crc kubenswrapper[4578]: E1003 13:27:47.651079 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac850868-8034-40e9-ac47-21455b1f8817" containerName="ssh-known-hosts-edpm-deployment" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.651103 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac850868-8034-40e9-ac47-21455b1f8817" containerName="ssh-known-hosts-edpm-deployment" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.651284 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac850868-8034-40e9-ac47-21455b1f8817" containerName="ssh-known-hosts-edpm-deployment" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.651890 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.653839 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-442mn" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.654060 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.658727 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.659853 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.670344 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46"] Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.709328 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/187edc5e-0ec9-4d8f-8085-2b0a393df59a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-nmz46\" (UID: \"187edc5e-0ec9-4d8f-8085-2b0a393df59a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.709451 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/187edc5e-0ec9-4d8f-8085-2b0a393df59a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-nmz46\" (UID: \"187edc5e-0ec9-4d8f-8085-2b0a393df59a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.709562 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgk79\" (UniqueName: \"kubernetes.io/projected/187edc5e-0ec9-4d8f-8085-2b0a393df59a-kube-api-access-cgk79\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-nmz46\" (UID: \"187edc5e-0ec9-4d8f-8085-2b0a393df59a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.810595 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/187edc5e-0ec9-4d8f-8085-2b0a393df59a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-nmz46\" (UID: \"187edc5e-0ec9-4d8f-8085-2b0a393df59a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.810717 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/187edc5e-0ec9-4d8f-8085-2b0a393df59a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-nmz46\" (UID: \"187edc5e-0ec9-4d8f-8085-2b0a393df59a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.810771 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgk79\" (UniqueName: \"kubernetes.io/projected/187edc5e-0ec9-4d8f-8085-2b0a393df59a-kube-api-access-cgk79\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-nmz46\" (UID: \"187edc5e-0ec9-4d8f-8085-2b0a393df59a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.815001 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/187edc5e-0ec9-4d8f-8085-2b0a393df59a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-nmz46\" (UID: \"187edc5e-0ec9-4d8f-8085-2b0a393df59a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.815061 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/187edc5e-0ec9-4d8f-8085-2b0a393df59a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-nmz46\" (UID: \"187edc5e-0ec9-4d8f-8085-2b0a393df59a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.828219 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgk79\" (UniqueName: \"kubernetes.io/projected/187edc5e-0ec9-4d8f-8085-2b0a393df59a-kube-api-access-cgk79\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-nmz46\" (UID: \"187edc5e-0ec9-4d8f-8085-2b0a393df59a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46" Oct 03 13:27:47 crc kubenswrapper[4578]: I1003 13:27:47.979450 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46" Oct 03 13:27:48 crc kubenswrapper[4578]: I1003 13:27:48.493705 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46"] Oct 03 13:27:48 crc kubenswrapper[4578]: I1003 13:27:48.581357 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46" event={"ID":"187edc5e-0ec9-4d8f-8085-2b0a393df59a","Type":"ContainerStarted","Data":"4ed29108ced5ad60dc7974110283b9df0f2cfac887b370beeff1128cda688a8f"} Oct 03 13:27:49 crc kubenswrapper[4578]: I1003 13:27:49.593911 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46" event={"ID":"187edc5e-0ec9-4d8f-8085-2b0a393df59a","Type":"ContainerStarted","Data":"c5be40f164f27a70923a9c0bbfc3c85f0e0ef0f5394ad28803056ef2f8fbdc78"} Oct 03 13:27:49 crc kubenswrapper[4578]: I1003 13:27:49.618198 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46" podStartSLOduration=2.452423084 podStartE2EDuration="2.618176914s" podCreationTimestamp="2025-10-03 13:27:47 +0000 UTC" firstStartedPulling="2025-10-03 13:27:48.505589162 +0000 UTC m=+2204.304061346" lastFinishedPulling="2025-10-03 13:27:48.671342992 +0000 UTC m=+2204.469815176" observedRunningTime="2025-10-03 13:27:49.60912153 +0000 UTC m=+2205.407593714" watchObservedRunningTime="2025-10-03 13:27:49.618176914 +0000 UTC m=+2205.416649118" Oct 03 13:27:50 crc kubenswrapper[4578]: I1003 13:27:50.908995 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:27:50 crc kubenswrapper[4578]: E1003 13:27:50.909833 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:27:57 crc kubenswrapper[4578]: I1003 13:27:57.659234 4578 generic.go:334] "Generic (PLEG): container finished" podID="187edc5e-0ec9-4d8f-8085-2b0a393df59a" containerID="c5be40f164f27a70923a9c0bbfc3c85f0e0ef0f5394ad28803056ef2f8fbdc78" exitCode=0 Oct 03 13:27:57 crc kubenswrapper[4578]: I1003 13:27:57.659320 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46" event={"ID":"187edc5e-0ec9-4d8f-8085-2b0a393df59a","Type":"ContainerDied","Data":"c5be40f164f27a70923a9c0bbfc3c85f0e0ef0f5394ad28803056ef2f8fbdc78"} Oct 03 13:27:58 crc kubenswrapper[4578]: I1003 13:27:58.406326 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8r884"] Oct 03 13:27:58 crc kubenswrapper[4578]: I1003 13:27:58.408965 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8r884" Oct 03 13:27:58 crc kubenswrapper[4578]: I1003 13:27:58.432331 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8r884"] Oct 03 13:27:58 crc kubenswrapper[4578]: I1003 13:27:58.518905 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15d01478-0ddc-4fd2-abd2-0f461d53492e-catalog-content\") pod \"redhat-operators-8r884\" (UID: \"15d01478-0ddc-4fd2-abd2-0f461d53492e\") " pod="openshift-marketplace/redhat-operators-8r884" Oct 03 13:27:58 crc kubenswrapper[4578]: I1003 13:27:58.519409 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15d01478-0ddc-4fd2-abd2-0f461d53492e-utilities\") pod \"redhat-operators-8r884\" (UID: \"15d01478-0ddc-4fd2-abd2-0f461d53492e\") " pod="openshift-marketplace/redhat-operators-8r884" Oct 03 13:27:58 crc kubenswrapper[4578]: I1003 13:27:58.519494 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpq2d\" (UniqueName: \"kubernetes.io/projected/15d01478-0ddc-4fd2-abd2-0f461d53492e-kube-api-access-kpq2d\") pod \"redhat-operators-8r884\" (UID: \"15d01478-0ddc-4fd2-abd2-0f461d53492e\") " pod="openshift-marketplace/redhat-operators-8r884" Oct 03 13:27:58 crc kubenswrapper[4578]: I1003 13:27:58.621460 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15d01478-0ddc-4fd2-abd2-0f461d53492e-catalog-content\") pod \"redhat-operators-8r884\" (UID: \"15d01478-0ddc-4fd2-abd2-0f461d53492e\") " pod="openshift-marketplace/redhat-operators-8r884" Oct 03 13:27:58 crc kubenswrapper[4578]: I1003 13:27:58.621526 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15d01478-0ddc-4fd2-abd2-0f461d53492e-utilities\") pod \"redhat-operators-8r884\" (UID: \"15d01478-0ddc-4fd2-abd2-0f461d53492e\") " pod="openshift-marketplace/redhat-operators-8r884" Oct 03 13:27:58 crc kubenswrapper[4578]: I1003 13:27:58.621547 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpq2d\" (UniqueName: \"kubernetes.io/projected/15d01478-0ddc-4fd2-abd2-0f461d53492e-kube-api-access-kpq2d\") pod \"redhat-operators-8r884\" (UID: \"15d01478-0ddc-4fd2-abd2-0f461d53492e\") " pod="openshift-marketplace/redhat-operators-8r884" Oct 03 13:27:58 crc kubenswrapper[4578]: I1003 13:27:58.622093 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15d01478-0ddc-4fd2-abd2-0f461d53492e-utilities\") pod \"redhat-operators-8r884\" (UID: \"15d01478-0ddc-4fd2-abd2-0f461d53492e\") " pod="openshift-marketplace/redhat-operators-8r884" Oct 03 13:27:58 crc kubenswrapper[4578]: I1003 13:27:58.622160 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15d01478-0ddc-4fd2-abd2-0f461d53492e-catalog-content\") pod \"redhat-operators-8r884\" (UID: \"15d01478-0ddc-4fd2-abd2-0f461d53492e\") " pod="openshift-marketplace/redhat-operators-8r884" Oct 03 13:27:58 crc kubenswrapper[4578]: I1003 13:27:58.645525 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpq2d\" (UniqueName: \"kubernetes.io/projected/15d01478-0ddc-4fd2-abd2-0f461d53492e-kube-api-access-kpq2d\") pod \"redhat-operators-8r884\" (UID: \"15d01478-0ddc-4fd2-abd2-0f461d53492e\") " pod="openshift-marketplace/redhat-operators-8r884" Oct 03 13:27:58 crc kubenswrapper[4578]: I1003 13:27:58.735602 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8r884" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.213778 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.317071 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8r884"] Oct 03 13:27:59 crc kubenswrapper[4578]: W1003 13:27:59.319204 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15d01478_0ddc_4fd2_abd2_0f461d53492e.slice/crio-894e9d7bcb528138e5fe4801f61972e52e631abb2a22ac6e81461c06cb894700 WatchSource:0}: Error finding container 894e9d7bcb528138e5fe4801f61972e52e631abb2a22ac6e81461c06cb894700: Status 404 returned error can't find the container with id 894e9d7bcb528138e5fe4801f61972e52e631abb2a22ac6e81461c06cb894700 Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.342378 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgk79\" (UniqueName: \"kubernetes.io/projected/187edc5e-0ec9-4d8f-8085-2b0a393df59a-kube-api-access-cgk79\") pod \"187edc5e-0ec9-4d8f-8085-2b0a393df59a\" (UID: \"187edc5e-0ec9-4d8f-8085-2b0a393df59a\") " Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.342462 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/187edc5e-0ec9-4d8f-8085-2b0a393df59a-ssh-key\") pod \"187edc5e-0ec9-4d8f-8085-2b0a393df59a\" (UID: \"187edc5e-0ec9-4d8f-8085-2b0a393df59a\") " Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.342530 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/187edc5e-0ec9-4d8f-8085-2b0a393df59a-inventory\") pod \"187edc5e-0ec9-4d8f-8085-2b0a393df59a\" (UID: \"187edc5e-0ec9-4d8f-8085-2b0a393df59a\") " Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.348855 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/187edc5e-0ec9-4d8f-8085-2b0a393df59a-kube-api-access-cgk79" (OuterVolumeSpecName: "kube-api-access-cgk79") pod "187edc5e-0ec9-4d8f-8085-2b0a393df59a" (UID: "187edc5e-0ec9-4d8f-8085-2b0a393df59a"). InnerVolumeSpecName "kube-api-access-cgk79". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.376582 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/187edc5e-0ec9-4d8f-8085-2b0a393df59a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "187edc5e-0ec9-4d8f-8085-2b0a393df59a" (UID: "187edc5e-0ec9-4d8f-8085-2b0a393df59a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.384779 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/187edc5e-0ec9-4d8f-8085-2b0a393df59a-inventory" (OuterVolumeSpecName: "inventory") pod "187edc5e-0ec9-4d8f-8085-2b0a393df59a" (UID: "187edc5e-0ec9-4d8f-8085-2b0a393df59a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.443836 4578 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/187edc5e-0ec9-4d8f-8085-2b0a393df59a-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.443864 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgk79\" (UniqueName: \"kubernetes.io/projected/187edc5e-0ec9-4d8f-8085-2b0a393df59a-kube-api-access-cgk79\") on node \"crc\" DevicePath \"\"" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.443874 4578 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/187edc5e-0ec9-4d8f-8085-2b0a393df59a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.683304 4578 generic.go:334] "Generic (PLEG): container finished" podID="15d01478-0ddc-4fd2-abd2-0f461d53492e" containerID="fb34bca9005475e9f4be3bd1cc26f3f766ba01ccfcfa87118dbc07b1a102f089" exitCode=0 Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.683354 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8r884" event={"ID":"15d01478-0ddc-4fd2-abd2-0f461d53492e","Type":"ContainerDied","Data":"fb34bca9005475e9f4be3bd1cc26f3f766ba01ccfcfa87118dbc07b1a102f089"} Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.683727 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8r884" event={"ID":"15d01478-0ddc-4fd2-abd2-0f461d53492e","Type":"ContainerStarted","Data":"894e9d7bcb528138e5fe4801f61972e52e631abb2a22ac6e81461c06cb894700"} Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.685668 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46" event={"ID":"187edc5e-0ec9-4d8f-8085-2b0a393df59a","Type":"ContainerDied","Data":"4ed29108ced5ad60dc7974110283b9df0f2cfac887b370beeff1128cda688a8f"} Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.685705 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ed29108ced5ad60dc7974110283b9df0f2cfac887b370beeff1128cda688a8f" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.685754 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-nmz46" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.805914 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8"] Oct 03 13:27:59 crc kubenswrapper[4578]: E1003 13:27:59.806287 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="187edc5e-0ec9-4d8f-8085-2b0a393df59a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.806301 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="187edc5e-0ec9-4d8f-8085-2b0a393df59a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.806469 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="187edc5e-0ec9-4d8f-8085-2b0a393df59a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.807088 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.808696 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.808863 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-442mn" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.809163 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.812940 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.822541 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8"] Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.869777 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-479g8\" (UID: \"5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.869856 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s6sj\" (UniqueName: \"kubernetes.io/projected/5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f-kube-api-access-9s6sj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-479g8\" (UID: \"5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.869909 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-479g8\" (UID: \"5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.971699 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-479g8\" (UID: \"5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.971813 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-479g8\" (UID: \"5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.971870 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s6sj\" (UniqueName: \"kubernetes.io/projected/5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f-kube-api-access-9s6sj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-479g8\" (UID: \"5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.977328 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-479g8\" (UID: \"5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.977672 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-479g8\" (UID: \"5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8" Oct 03 13:27:59 crc kubenswrapper[4578]: I1003 13:27:59.996602 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s6sj\" (UniqueName: \"kubernetes.io/projected/5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f-kube-api-access-9s6sj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-479g8\" (UID: \"5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8" Oct 03 13:28:00 crc kubenswrapper[4578]: I1003 13:28:00.127577 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8" Oct 03 13:28:00 crc kubenswrapper[4578]: W1003 13:28:00.677259 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d79d7a8_f4ba_4f27_8860_97c9d8f36b8f.slice/crio-11a50e2f2b5700f837a29a871ea3a8357eac7823494f2e62984a378b8cfb7100 WatchSource:0}: Error finding container 11a50e2f2b5700f837a29a871ea3a8357eac7823494f2e62984a378b8cfb7100: Status 404 returned error can't find the container with id 11a50e2f2b5700f837a29a871ea3a8357eac7823494f2e62984a378b8cfb7100 Oct 03 13:28:00 crc kubenswrapper[4578]: I1003 13:28:00.685975 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8"] Oct 03 13:28:00 crc kubenswrapper[4578]: I1003 13:28:00.696460 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8" event={"ID":"5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f","Type":"ContainerStarted","Data":"11a50e2f2b5700f837a29a871ea3a8357eac7823494f2e62984a378b8cfb7100"} Oct 03 13:28:01 crc kubenswrapper[4578]: I1003 13:28:01.708016 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8r884" event={"ID":"15d01478-0ddc-4fd2-abd2-0f461d53492e","Type":"ContainerStarted","Data":"1dc44dd3ec0da4f12a4036251b6cb8e12c31546e157ed9e0747d6f995bdd6a52"} Oct 03 13:28:01 crc kubenswrapper[4578]: I1003 13:28:01.711016 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8" event={"ID":"5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f","Type":"ContainerStarted","Data":"5e44af1400a99b3fe886d6203c6f0340752639a8ab41b85a4035b1ced1a341c6"} Oct 03 13:28:01 crc kubenswrapper[4578]: I1003 13:28:01.786542 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8" podStartSLOduration=2.415632124 podStartE2EDuration="2.786523533s" podCreationTimestamp="2025-10-03 13:27:59 +0000 UTC" firstStartedPulling="2025-10-03 13:28:00.678793242 +0000 UTC m=+2216.477265416" lastFinishedPulling="2025-10-03 13:28:01.049684641 +0000 UTC m=+2216.848156825" observedRunningTime="2025-10-03 13:28:01.763084637 +0000 UTC m=+2217.561556841" watchObservedRunningTime="2025-10-03 13:28:01.786523533 +0000 UTC m=+2217.584995707" Oct 03 13:28:02 crc kubenswrapper[4578]: I1003 13:28:02.909792 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:28:02 crc kubenswrapper[4578]: E1003 13:28:02.910447 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:28:08 crc kubenswrapper[4578]: I1003 13:28:08.779921 4578 generic.go:334] "Generic (PLEG): container finished" podID="15d01478-0ddc-4fd2-abd2-0f461d53492e" containerID="1dc44dd3ec0da4f12a4036251b6cb8e12c31546e157ed9e0747d6f995bdd6a52" exitCode=0 Oct 03 13:28:08 crc kubenswrapper[4578]: I1003 13:28:08.780011 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8r884" event={"ID":"15d01478-0ddc-4fd2-abd2-0f461d53492e","Type":"ContainerDied","Data":"1dc44dd3ec0da4f12a4036251b6cb8e12c31546e157ed9e0747d6f995bdd6a52"} Oct 03 13:28:09 crc kubenswrapper[4578]: I1003 13:28:09.790835 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8r884" event={"ID":"15d01478-0ddc-4fd2-abd2-0f461d53492e","Type":"ContainerStarted","Data":"e35e3525f9ed3e04e90739c0299cf4d3dbfab224a381408758b214fa635a34da"} Oct 03 13:28:09 crc kubenswrapper[4578]: I1003 13:28:09.823676 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8r884" podStartSLOduration=2.349850772 podStartE2EDuration="11.823655157s" podCreationTimestamp="2025-10-03 13:27:58 +0000 UTC" firstStartedPulling="2025-10-03 13:27:59.685281787 +0000 UTC m=+2215.483753971" lastFinishedPulling="2025-10-03 13:28:09.159086172 +0000 UTC m=+2224.957558356" observedRunningTime="2025-10-03 13:28:09.816207143 +0000 UTC m=+2225.614679327" watchObservedRunningTime="2025-10-03 13:28:09.823655157 +0000 UTC m=+2225.622127341" Oct 03 13:28:11 crc kubenswrapper[4578]: I1003 13:28:11.808603 4578 generic.go:334] "Generic (PLEG): container finished" podID="5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f" containerID="5e44af1400a99b3fe886d6203c6f0340752639a8ab41b85a4035b1ced1a341c6" exitCode=0 Oct 03 13:28:11 crc kubenswrapper[4578]: I1003 13:28:11.808671 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8" event={"ID":"5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f","Type":"ContainerDied","Data":"5e44af1400a99b3fe886d6203c6f0340752639a8ab41b85a4035b1ced1a341c6"} Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.229836 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8" Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.320987 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s6sj\" (UniqueName: \"kubernetes.io/projected/5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f-kube-api-access-9s6sj\") pod \"5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f\" (UID: \"5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f\") " Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.321084 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f-inventory\") pod \"5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f\" (UID: \"5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f\") " Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.321118 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f-ssh-key\") pod \"5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f\" (UID: \"5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f\") " Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.334338 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f-kube-api-access-9s6sj" (OuterVolumeSpecName: "kube-api-access-9s6sj") pod "5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f" (UID: "5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f"). InnerVolumeSpecName "kube-api-access-9s6sj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.355144 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f" (UID: "5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.355804 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f-inventory" (OuterVolumeSpecName: "inventory") pod "5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f" (UID: "5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.423430 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9s6sj\" (UniqueName: \"kubernetes.io/projected/5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f-kube-api-access-9s6sj\") on node \"crc\" DevicePath \"\"" Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.423458 4578 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.423467 4578 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.827990 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8" event={"ID":"5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f","Type":"ContainerDied","Data":"11a50e2f2b5700f837a29a871ea3a8357eac7823494f2e62984a378b8cfb7100"} Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.828028 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11a50e2f2b5700f837a29a871ea3a8357eac7823494f2e62984a378b8cfb7100" Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.828083 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-479g8" Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.946111 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp"] Oct 03 13:28:13 crc kubenswrapper[4578]: E1003 13:28:13.946828 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.946842 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.947097 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.947788 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.981443 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.983187 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.983611 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.986324 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-442mn" Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.987853 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.991798 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.992343 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 03 13:28:13 crc kubenswrapper[4578]: I1003 13:28:13.999696 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.019654 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp"] Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.042422 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.042498 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.042538 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.042592 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.042669 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.042901 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.043004 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.043089 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.043161 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4ztw\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-kube-api-access-t4ztw\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.043249 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.043333 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.043485 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.043586 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.043690 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.145336 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.145393 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4ztw\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-kube-api-access-t4ztw\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.145440 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.145473 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.145524 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.145566 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.145599 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.145700 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.145755 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.145785 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.145838 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.145886 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.146264 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.146309 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.151948 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.163485 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.176926 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.177027 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.177254 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.177611 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.178414 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.179053 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.179402 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.179853 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.181780 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.182729 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.182984 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.191474 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4ztw\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-kube-api-access-t4ztw\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trqwp\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.320736 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:14 crc kubenswrapper[4578]: I1003 13:28:14.878272 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp"] Oct 03 13:28:14 crc kubenswrapper[4578]: W1003 13:28:14.892738 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76db17db_5757_4c06_94d4_537c645f54e5.slice/crio-008922388fd7446d7a20545848f08dd679e473ac5b5b451b15b9d764fefc685b WatchSource:0}: Error finding container 008922388fd7446d7a20545848f08dd679e473ac5b5b451b15b9d764fefc685b: Status 404 returned error can't find the container with id 008922388fd7446d7a20545848f08dd679e473ac5b5b451b15b9d764fefc685b Oct 03 13:28:15 crc kubenswrapper[4578]: I1003 13:28:15.844320 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" event={"ID":"76db17db-5757-4c06-94d4-537c645f54e5","Type":"ContainerStarted","Data":"4d3b6e405be7d36441858474e9d8cc95f11c16782494bbb164c34aee6726d14f"} Oct 03 13:28:15 crc kubenswrapper[4578]: I1003 13:28:15.844685 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" event={"ID":"76db17db-5757-4c06-94d4-537c645f54e5","Type":"ContainerStarted","Data":"008922388fd7446d7a20545848f08dd679e473ac5b5b451b15b9d764fefc685b"} Oct 03 13:28:15 crc kubenswrapper[4578]: I1003 13:28:15.880478 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" podStartSLOduration=2.466546509 podStartE2EDuration="2.880459897s" podCreationTimestamp="2025-10-03 13:28:13 +0000 UTC" firstStartedPulling="2025-10-03 13:28:14.894349254 +0000 UTC m=+2230.692821438" lastFinishedPulling="2025-10-03 13:28:15.308262642 +0000 UTC m=+2231.106734826" observedRunningTime="2025-10-03 13:28:15.870146564 +0000 UTC m=+2231.668618748" watchObservedRunningTime="2025-10-03 13:28:15.880459897 +0000 UTC m=+2231.678932081" Oct 03 13:28:16 crc kubenswrapper[4578]: I1003 13:28:16.908763 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:28:16 crc kubenswrapper[4578]: E1003 13:28:16.909017 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:28:18 crc kubenswrapper[4578]: I1003 13:28:18.736643 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8r884" Oct 03 13:28:18 crc kubenswrapper[4578]: I1003 13:28:18.739159 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8r884" Oct 03 13:28:19 crc kubenswrapper[4578]: I1003 13:28:19.780299 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8r884" podUID="15d01478-0ddc-4fd2-abd2-0f461d53492e" containerName="registry-server" probeResult="failure" output=< Oct 03 13:28:19 crc kubenswrapper[4578]: timeout: failed to connect service ":50051" within 1s Oct 03 13:28:19 crc kubenswrapper[4578]: > Oct 03 13:28:29 crc kubenswrapper[4578]: I1003 13:28:29.786191 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8r884" podUID="15d01478-0ddc-4fd2-abd2-0f461d53492e" containerName="registry-server" probeResult="failure" output=< Oct 03 13:28:29 crc kubenswrapper[4578]: timeout: failed to connect service ":50051" within 1s Oct 03 13:28:29 crc kubenswrapper[4578]: > Oct 03 13:28:31 crc kubenswrapper[4578]: I1003 13:28:31.909506 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:28:31 crc kubenswrapper[4578]: E1003 13:28:31.910104 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:28:38 crc kubenswrapper[4578]: I1003 13:28:38.782140 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8r884" Oct 03 13:28:38 crc kubenswrapper[4578]: I1003 13:28:38.836269 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8r884" Oct 03 13:28:39 crc kubenswrapper[4578]: I1003 13:28:39.015254 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8r884"] Oct 03 13:28:40 crc kubenswrapper[4578]: I1003 13:28:40.064243 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8r884" podUID="15d01478-0ddc-4fd2-abd2-0f461d53492e" containerName="registry-server" containerID="cri-o://e35e3525f9ed3e04e90739c0299cf4d3dbfab224a381408758b214fa635a34da" gracePeriod=2 Oct 03 13:28:40 crc kubenswrapper[4578]: I1003 13:28:40.559582 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8r884" Oct 03 13:28:40 crc kubenswrapper[4578]: I1003 13:28:40.683870 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15d01478-0ddc-4fd2-abd2-0f461d53492e-catalog-content\") pod \"15d01478-0ddc-4fd2-abd2-0f461d53492e\" (UID: \"15d01478-0ddc-4fd2-abd2-0f461d53492e\") " Oct 03 13:28:40 crc kubenswrapper[4578]: I1003 13:28:40.686868 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpq2d\" (UniqueName: \"kubernetes.io/projected/15d01478-0ddc-4fd2-abd2-0f461d53492e-kube-api-access-kpq2d\") pod \"15d01478-0ddc-4fd2-abd2-0f461d53492e\" (UID: \"15d01478-0ddc-4fd2-abd2-0f461d53492e\") " Oct 03 13:28:40 crc kubenswrapper[4578]: I1003 13:28:40.687111 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15d01478-0ddc-4fd2-abd2-0f461d53492e-utilities\") pod \"15d01478-0ddc-4fd2-abd2-0f461d53492e\" (UID: \"15d01478-0ddc-4fd2-abd2-0f461d53492e\") " Oct 03 13:28:40 crc kubenswrapper[4578]: I1003 13:28:40.690430 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15d01478-0ddc-4fd2-abd2-0f461d53492e-utilities" (OuterVolumeSpecName: "utilities") pod "15d01478-0ddc-4fd2-abd2-0f461d53492e" (UID: "15d01478-0ddc-4fd2-abd2-0f461d53492e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:28:40 crc kubenswrapper[4578]: I1003 13:28:40.725426 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15d01478-0ddc-4fd2-abd2-0f461d53492e-kube-api-access-kpq2d" (OuterVolumeSpecName: "kube-api-access-kpq2d") pod "15d01478-0ddc-4fd2-abd2-0f461d53492e" (UID: "15d01478-0ddc-4fd2-abd2-0f461d53492e"). InnerVolumeSpecName "kube-api-access-kpq2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:28:40 crc kubenswrapper[4578]: I1003 13:28:40.786496 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15d01478-0ddc-4fd2-abd2-0f461d53492e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15d01478-0ddc-4fd2-abd2-0f461d53492e" (UID: "15d01478-0ddc-4fd2-abd2-0f461d53492e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:28:40 crc kubenswrapper[4578]: I1003 13:28:40.790621 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpq2d\" (UniqueName: \"kubernetes.io/projected/15d01478-0ddc-4fd2-abd2-0f461d53492e-kube-api-access-kpq2d\") on node \"crc\" DevicePath \"\"" Oct 03 13:28:40 crc kubenswrapper[4578]: I1003 13:28:40.790673 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15d01478-0ddc-4fd2-abd2-0f461d53492e-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:28:40 crc kubenswrapper[4578]: I1003 13:28:40.790685 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15d01478-0ddc-4fd2-abd2-0f461d53492e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:28:41 crc kubenswrapper[4578]: I1003 13:28:41.074598 4578 generic.go:334] "Generic (PLEG): container finished" podID="15d01478-0ddc-4fd2-abd2-0f461d53492e" containerID="e35e3525f9ed3e04e90739c0299cf4d3dbfab224a381408758b214fa635a34da" exitCode=0 Oct 03 13:28:41 crc kubenswrapper[4578]: I1003 13:28:41.074670 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8r884" event={"ID":"15d01478-0ddc-4fd2-abd2-0f461d53492e","Type":"ContainerDied","Data":"e35e3525f9ed3e04e90739c0299cf4d3dbfab224a381408758b214fa635a34da"} Oct 03 13:28:41 crc kubenswrapper[4578]: I1003 13:28:41.074709 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8r884" Oct 03 13:28:41 crc kubenswrapper[4578]: I1003 13:28:41.074743 4578 scope.go:117] "RemoveContainer" containerID="e35e3525f9ed3e04e90739c0299cf4d3dbfab224a381408758b214fa635a34da" Oct 03 13:28:41 crc kubenswrapper[4578]: I1003 13:28:41.074729 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8r884" event={"ID":"15d01478-0ddc-4fd2-abd2-0f461d53492e","Type":"ContainerDied","Data":"894e9d7bcb528138e5fe4801f61972e52e631abb2a22ac6e81461c06cb894700"} Oct 03 13:28:41 crc kubenswrapper[4578]: I1003 13:28:41.099880 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8r884"] Oct 03 13:28:41 crc kubenswrapper[4578]: I1003 13:28:41.103102 4578 scope.go:117] "RemoveContainer" containerID="1dc44dd3ec0da4f12a4036251b6cb8e12c31546e157ed9e0747d6f995bdd6a52" Oct 03 13:28:41 crc kubenswrapper[4578]: I1003 13:28:41.108610 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8r884"] Oct 03 13:28:41 crc kubenswrapper[4578]: I1003 13:28:41.128806 4578 scope.go:117] "RemoveContainer" containerID="fb34bca9005475e9f4be3bd1cc26f3f766ba01ccfcfa87118dbc07b1a102f089" Oct 03 13:28:41 crc kubenswrapper[4578]: I1003 13:28:41.185316 4578 scope.go:117] "RemoveContainer" containerID="e35e3525f9ed3e04e90739c0299cf4d3dbfab224a381408758b214fa635a34da" Oct 03 13:28:41 crc kubenswrapper[4578]: E1003 13:28:41.185971 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e35e3525f9ed3e04e90739c0299cf4d3dbfab224a381408758b214fa635a34da\": container with ID starting with e35e3525f9ed3e04e90739c0299cf4d3dbfab224a381408758b214fa635a34da not found: ID does not exist" containerID="e35e3525f9ed3e04e90739c0299cf4d3dbfab224a381408758b214fa635a34da" Oct 03 13:28:41 crc kubenswrapper[4578]: I1003 13:28:41.186006 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e35e3525f9ed3e04e90739c0299cf4d3dbfab224a381408758b214fa635a34da"} err="failed to get container status \"e35e3525f9ed3e04e90739c0299cf4d3dbfab224a381408758b214fa635a34da\": rpc error: code = NotFound desc = could not find container \"e35e3525f9ed3e04e90739c0299cf4d3dbfab224a381408758b214fa635a34da\": container with ID starting with e35e3525f9ed3e04e90739c0299cf4d3dbfab224a381408758b214fa635a34da not found: ID does not exist" Oct 03 13:28:41 crc kubenswrapper[4578]: I1003 13:28:41.186048 4578 scope.go:117] "RemoveContainer" containerID="1dc44dd3ec0da4f12a4036251b6cb8e12c31546e157ed9e0747d6f995bdd6a52" Oct 03 13:28:41 crc kubenswrapper[4578]: E1003 13:28:41.186338 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dc44dd3ec0da4f12a4036251b6cb8e12c31546e157ed9e0747d6f995bdd6a52\": container with ID starting with 1dc44dd3ec0da4f12a4036251b6cb8e12c31546e157ed9e0747d6f995bdd6a52 not found: ID does not exist" containerID="1dc44dd3ec0da4f12a4036251b6cb8e12c31546e157ed9e0747d6f995bdd6a52" Oct 03 13:28:41 crc kubenswrapper[4578]: I1003 13:28:41.186363 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dc44dd3ec0da4f12a4036251b6cb8e12c31546e157ed9e0747d6f995bdd6a52"} err="failed to get container status \"1dc44dd3ec0da4f12a4036251b6cb8e12c31546e157ed9e0747d6f995bdd6a52\": rpc error: code = NotFound desc = could not find container \"1dc44dd3ec0da4f12a4036251b6cb8e12c31546e157ed9e0747d6f995bdd6a52\": container with ID starting with 1dc44dd3ec0da4f12a4036251b6cb8e12c31546e157ed9e0747d6f995bdd6a52 not found: ID does not exist" Oct 03 13:28:41 crc kubenswrapper[4578]: I1003 13:28:41.186377 4578 scope.go:117] "RemoveContainer" containerID="fb34bca9005475e9f4be3bd1cc26f3f766ba01ccfcfa87118dbc07b1a102f089" Oct 03 13:28:41 crc kubenswrapper[4578]: E1003 13:28:41.186595 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb34bca9005475e9f4be3bd1cc26f3f766ba01ccfcfa87118dbc07b1a102f089\": container with ID starting with fb34bca9005475e9f4be3bd1cc26f3f766ba01ccfcfa87118dbc07b1a102f089 not found: ID does not exist" containerID="fb34bca9005475e9f4be3bd1cc26f3f766ba01ccfcfa87118dbc07b1a102f089" Oct 03 13:28:41 crc kubenswrapper[4578]: I1003 13:28:41.186618 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb34bca9005475e9f4be3bd1cc26f3f766ba01ccfcfa87118dbc07b1a102f089"} err="failed to get container status \"fb34bca9005475e9f4be3bd1cc26f3f766ba01ccfcfa87118dbc07b1a102f089\": rpc error: code = NotFound desc = could not find container \"fb34bca9005475e9f4be3bd1cc26f3f766ba01ccfcfa87118dbc07b1a102f089\": container with ID starting with fb34bca9005475e9f4be3bd1cc26f3f766ba01ccfcfa87118dbc07b1a102f089 not found: ID does not exist" Oct 03 13:28:42 crc kubenswrapper[4578]: I1003 13:28:42.908822 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:28:42 crc kubenswrapper[4578]: E1003 13:28:42.909443 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:28:42 crc kubenswrapper[4578]: I1003 13:28:42.920574 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15d01478-0ddc-4fd2-abd2-0f461d53492e" path="/var/lib/kubelet/pods/15d01478-0ddc-4fd2-abd2-0f461d53492e/volumes" Oct 03 13:28:54 crc kubenswrapper[4578]: E1003 13:28:54.619060 4578 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76db17db_5757_4c06_94d4_537c645f54e5.slice/crio-conmon-4d3b6e405be7d36441858474e9d8cc95f11c16782494bbb164c34aee6726d14f.scope\": RecentStats: unable to find data in memory cache]" Oct 03 13:28:55 crc kubenswrapper[4578]: I1003 13:28:55.192934 4578 generic.go:334] "Generic (PLEG): container finished" podID="76db17db-5757-4c06-94d4-537c645f54e5" containerID="4d3b6e405be7d36441858474e9d8cc95f11c16782494bbb164c34aee6726d14f" exitCode=0 Oct 03 13:28:55 crc kubenswrapper[4578]: I1003 13:28:55.193296 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" event={"ID":"76db17db-5757-4c06-94d4-537c645f54e5","Type":"ContainerDied","Data":"4d3b6e405be7d36441858474e9d8cc95f11c16782494bbb164c34aee6726d14f"} Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.610571 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.780100 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"76db17db-5757-4c06-94d4-537c645f54e5\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.780169 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-telemetry-combined-ca-bundle\") pod \"76db17db-5757-4c06-94d4-537c645f54e5\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.780212 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-bootstrap-combined-ca-bundle\") pod \"76db17db-5757-4c06-94d4-537c645f54e5\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.780232 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-ovn-combined-ca-bundle\") pod \"76db17db-5757-4c06-94d4-537c645f54e5\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.780275 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-inventory\") pod \"76db17db-5757-4c06-94d4-537c645f54e5\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.780300 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-ssh-key\") pod \"76db17db-5757-4c06-94d4-537c645f54e5\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.780321 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-ovn-default-certs-0\") pod \"76db17db-5757-4c06-94d4-537c645f54e5\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.780348 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-nova-combined-ca-bundle\") pod \"76db17db-5757-4c06-94d4-537c645f54e5\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.780376 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"76db17db-5757-4c06-94d4-537c645f54e5\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.780411 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-neutron-metadata-combined-ca-bundle\") pod \"76db17db-5757-4c06-94d4-537c645f54e5\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.780430 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"76db17db-5757-4c06-94d4-537c645f54e5\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.780467 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-libvirt-combined-ca-bundle\") pod \"76db17db-5757-4c06-94d4-537c645f54e5\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.780484 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-repo-setup-combined-ca-bundle\") pod \"76db17db-5757-4c06-94d4-537c645f54e5\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.780557 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4ztw\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-kube-api-access-t4ztw\") pod \"76db17db-5757-4c06-94d4-537c645f54e5\" (UID: \"76db17db-5757-4c06-94d4-537c645f54e5\") " Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.787564 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-kube-api-access-t4ztw" (OuterVolumeSpecName: "kube-api-access-t4ztw") pod "76db17db-5757-4c06-94d4-537c645f54e5" (UID: "76db17db-5757-4c06-94d4-537c645f54e5"). InnerVolumeSpecName "kube-api-access-t4ztw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.789369 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "76db17db-5757-4c06-94d4-537c645f54e5" (UID: "76db17db-5757-4c06-94d4-537c645f54e5"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.789379 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "76db17db-5757-4c06-94d4-537c645f54e5" (UID: "76db17db-5757-4c06-94d4-537c645f54e5"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.789470 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "76db17db-5757-4c06-94d4-537c645f54e5" (UID: "76db17db-5757-4c06-94d4-537c645f54e5"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.789514 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "76db17db-5757-4c06-94d4-537c645f54e5" (UID: "76db17db-5757-4c06-94d4-537c645f54e5"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.789595 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "76db17db-5757-4c06-94d4-537c645f54e5" (UID: "76db17db-5757-4c06-94d4-537c645f54e5"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.790400 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "76db17db-5757-4c06-94d4-537c645f54e5" (UID: "76db17db-5757-4c06-94d4-537c645f54e5"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.792102 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "76db17db-5757-4c06-94d4-537c645f54e5" (UID: "76db17db-5757-4c06-94d4-537c645f54e5"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.792545 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "76db17db-5757-4c06-94d4-537c645f54e5" (UID: "76db17db-5757-4c06-94d4-537c645f54e5"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.793017 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "76db17db-5757-4c06-94d4-537c645f54e5" (UID: "76db17db-5757-4c06-94d4-537c645f54e5"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.793477 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "76db17db-5757-4c06-94d4-537c645f54e5" (UID: "76db17db-5757-4c06-94d4-537c645f54e5"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.798288 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "76db17db-5757-4c06-94d4-537c645f54e5" (UID: "76db17db-5757-4c06-94d4-537c645f54e5"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.814871 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "76db17db-5757-4c06-94d4-537c645f54e5" (UID: "76db17db-5757-4c06-94d4-537c645f54e5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.814971 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-inventory" (OuterVolumeSpecName: "inventory") pod "76db17db-5757-4c06-94d4-537c645f54e5" (UID: "76db17db-5757-4c06-94d4-537c645f54e5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.883685 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4ztw\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-kube-api-access-t4ztw\") on node \"crc\" DevicePath \"\"" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.884065 4578 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.884082 4578 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.884097 4578 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.884109 4578 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.884121 4578 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.884131 4578 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.884142 4578 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.884154 4578 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.884166 4578 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.884179 4578 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.884190 4578 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/76db17db-5757-4c06-94d4-537c645f54e5-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.884202 4578 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.884214 4578 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db17db-5757-4c06-94d4-537c645f54e5-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:28:56 crc kubenswrapper[4578]: I1003 13:28:56.909719 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:28:56 crc kubenswrapper[4578]: E1003 13:28:56.910220 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.210088 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" event={"ID":"76db17db-5757-4c06-94d4-537c645f54e5","Type":"ContainerDied","Data":"008922388fd7446d7a20545848f08dd679e473ac5b5b451b15b9d764fefc685b"} Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.210347 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="008922388fd7446d7a20545848f08dd679e473ac5b5b451b15b9d764fefc685b" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.210148 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trqwp" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.345494 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b"] Oct 03 13:28:57 crc kubenswrapper[4578]: E1003 13:28:57.345901 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15d01478-0ddc-4fd2-abd2-0f461d53492e" containerName="extract-content" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.345919 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="15d01478-0ddc-4fd2-abd2-0f461d53492e" containerName="extract-content" Oct 03 13:28:57 crc kubenswrapper[4578]: E1003 13:28:57.345929 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76db17db-5757-4c06-94d4-537c645f54e5" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.345938 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="76db17db-5757-4c06-94d4-537c645f54e5" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 03 13:28:57 crc kubenswrapper[4578]: E1003 13:28:57.345965 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15d01478-0ddc-4fd2-abd2-0f461d53492e" containerName="registry-server" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.345971 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="15d01478-0ddc-4fd2-abd2-0f461d53492e" containerName="registry-server" Oct 03 13:28:57 crc kubenswrapper[4578]: E1003 13:28:57.345982 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15d01478-0ddc-4fd2-abd2-0f461d53492e" containerName="extract-utilities" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.345988 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="15d01478-0ddc-4fd2-abd2-0f461d53492e" containerName="extract-utilities" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.346151 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="15d01478-0ddc-4fd2-abd2-0f461d53492e" containerName="registry-server" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.346167 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="76db17db-5757-4c06-94d4-537c645f54e5" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.346784 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.350988 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.351258 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.351334 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-442mn" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.351489 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.353305 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.371850 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b"] Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.495317 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee86abf-8757-4c4e-8841-37d978cff99a-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tnz6b\" (UID: \"2ee86abf-8757-4c4e-8841-37d978cff99a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.495706 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ee86abf-8757-4c4e-8841-37d978cff99a-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tnz6b\" (UID: \"2ee86abf-8757-4c4e-8841-37d978cff99a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.496007 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/2ee86abf-8757-4c4e-8841-37d978cff99a-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tnz6b\" (UID: \"2ee86abf-8757-4c4e-8841-37d978cff99a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.496290 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx4xc\" (UniqueName: \"kubernetes.io/projected/2ee86abf-8757-4c4e-8841-37d978cff99a-kube-api-access-qx4xc\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tnz6b\" (UID: \"2ee86abf-8757-4c4e-8841-37d978cff99a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.496858 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ee86abf-8757-4c4e-8841-37d978cff99a-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tnz6b\" (UID: \"2ee86abf-8757-4c4e-8841-37d978cff99a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.598798 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx4xc\" (UniqueName: \"kubernetes.io/projected/2ee86abf-8757-4c4e-8841-37d978cff99a-kube-api-access-qx4xc\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tnz6b\" (UID: \"2ee86abf-8757-4c4e-8841-37d978cff99a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.599149 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ee86abf-8757-4c4e-8841-37d978cff99a-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tnz6b\" (UID: \"2ee86abf-8757-4c4e-8841-37d978cff99a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.599297 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee86abf-8757-4c4e-8841-37d978cff99a-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tnz6b\" (UID: \"2ee86abf-8757-4c4e-8841-37d978cff99a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.599451 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ee86abf-8757-4c4e-8841-37d978cff99a-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tnz6b\" (UID: \"2ee86abf-8757-4c4e-8841-37d978cff99a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.599566 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/2ee86abf-8757-4c4e-8841-37d978cff99a-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tnz6b\" (UID: \"2ee86abf-8757-4c4e-8841-37d978cff99a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.600531 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/2ee86abf-8757-4c4e-8841-37d978cff99a-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tnz6b\" (UID: \"2ee86abf-8757-4c4e-8841-37d978cff99a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.603362 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ee86abf-8757-4c4e-8841-37d978cff99a-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tnz6b\" (UID: \"2ee86abf-8757-4c4e-8841-37d978cff99a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.603577 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee86abf-8757-4c4e-8841-37d978cff99a-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tnz6b\" (UID: \"2ee86abf-8757-4c4e-8841-37d978cff99a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.604258 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ee86abf-8757-4c4e-8841-37d978cff99a-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tnz6b\" (UID: \"2ee86abf-8757-4c4e-8841-37d978cff99a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.619643 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx4xc\" (UniqueName: \"kubernetes.io/projected/2ee86abf-8757-4c4e-8841-37d978cff99a-kube-api-access-qx4xc\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-tnz6b\" (UID: \"2ee86abf-8757-4c4e-8841-37d978cff99a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" Oct 03 13:28:57 crc kubenswrapper[4578]: I1003 13:28:57.669275 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" Oct 03 13:28:58 crc kubenswrapper[4578]: I1003 13:28:58.170531 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b"] Oct 03 13:28:58 crc kubenswrapper[4578]: W1003 13:28:58.175617 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ee86abf_8757_4c4e_8841_37d978cff99a.slice/crio-2faaae5f732f849c8be55a995044f7009ba0e92dd1bb13e72f46c0e25305cb52 WatchSource:0}: Error finding container 2faaae5f732f849c8be55a995044f7009ba0e92dd1bb13e72f46c0e25305cb52: Status 404 returned error can't find the container with id 2faaae5f732f849c8be55a995044f7009ba0e92dd1bb13e72f46c0e25305cb52 Oct 03 13:28:58 crc kubenswrapper[4578]: I1003 13:28:58.218293 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" event={"ID":"2ee86abf-8757-4c4e-8841-37d978cff99a","Type":"ContainerStarted","Data":"2faaae5f732f849c8be55a995044f7009ba0e92dd1bb13e72f46c0e25305cb52"} Oct 03 13:28:59 crc kubenswrapper[4578]: I1003 13:28:59.229309 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" event={"ID":"2ee86abf-8757-4c4e-8841-37d978cff99a","Type":"ContainerStarted","Data":"ae22ae44a2db64aebcfe1f260e86fe189f6a0d0ec816d759a5f08198e1f7bc33"} Oct 03 13:28:59 crc kubenswrapper[4578]: I1003 13:28:59.254948 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" podStartSLOduration=1.962048767 podStartE2EDuration="2.254928626s" podCreationTimestamp="2025-10-03 13:28:57 +0000 UTC" firstStartedPulling="2025-10-03 13:28:58.177573989 +0000 UTC m=+2273.976046173" lastFinishedPulling="2025-10-03 13:28:58.470453848 +0000 UTC m=+2274.268926032" observedRunningTime="2025-10-03 13:28:59.24872073 +0000 UTC m=+2275.047192914" watchObservedRunningTime="2025-10-03 13:28:59.254928626 +0000 UTC m=+2275.053400810" Oct 03 13:29:10 crc kubenswrapper[4578]: I1003 13:29:10.909867 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:29:10 crc kubenswrapper[4578]: E1003 13:29:10.910934 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:29:25 crc kubenswrapper[4578]: I1003 13:29:25.909832 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:29:25 crc kubenswrapper[4578]: E1003 13:29:25.910805 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:29:32 crc kubenswrapper[4578]: I1003 13:29:32.518421 4578 generic.go:334] "Generic (PLEG): container finished" podID="2ee86abf-8757-4c4e-8841-37d978cff99a" containerID="ae22ae44a2db64aebcfe1f260e86fe189f6a0d0ec816d759a5f08198e1f7bc33" exitCode=2 Oct 03 13:29:32 crc kubenswrapper[4578]: I1003 13:29:32.518509 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" event={"ID":"2ee86abf-8757-4c4e-8841-37d978cff99a","Type":"ContainerDied","Data":"ae22ae44a2db64aebcfe1f260e86fe189f6a0d0ec816d759a5f08198e1f7bc33"} Oct 03 13:29:33 crc kubenswrapper[4578]: I1003 13:29:33.907104 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" Oct 03 13:29:34 crc kubenswrapper[4578]: I1003 13:29:34.074906 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ee86abf-8757-4c4e-8841-37d978cff99a-ssh-key\") pod \"2ee86abf-8757-4c4e-8841-37d978cff99a\" (UID: \"2ee86abf-8757-4c4e-8841-37d978cff99a\") " Oct 03 13:29:34 crc kubenswrapper[4578]: I1003 13:29:34.074944 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ee86abf-8757-4c4e-8841-37d978cff99a-inventory\") pod \"2ee86abf-8757-4c4e-8841-37d978cff99a\" (UID: \"2ee86abf-8757-4c4e-8841-37d978cff99a\") " Oct 03 13:29:34 crc kubenswrapper[4578]: I1003 13:29:34.075089 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qx4xc\" (UniqueName: \"kubernetes.io/projected/2ee86abf-8757-4c4e-8841-37d978cff99a-kube-api-access-qx4xc\") pod \"2ee86abf-8757-4c4e-8841-37d978cff99a\" (UID: \"2ee86abf-8757-4c4e-8841-37d978cff99a\") " Oct 03 13:29:34 crc kubenswrapper[4578]: I1003 13:29:34.075122 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee86abf-8757-4c4e-8841-37d978cff99a-ovn-combined-ca-bundle\") pod \"2ee86abf-8757-4c4e-8841-37d978cff99a\" (UID: \"2ee86abf-8757-4c4e-8841-37d978cff99a\") " Oct 03 13:29:34 crc kubenswrapper[4578]: I1003 13:29:34.075187 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/2ee86abf-8757-4c4e-8841-37d978cff99a-ovncontroller-config-0\") pod \"2ee86abf-8757-4c4e-8841-37d978cff99a\" (UID: \"2ee86abf-8757-4c4e-8841-37d978cff99a\") " Oct 03 13:29:34 crc kubenswrapper[4578]: I1003 13:29:34.080718 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ee86abf-8757-4c4e-8841-37d978cff99a-kube-api-access-qx4xc" (OuterVolumeSpecName: "kube-api-access-qx4xc") pod "2ee86abf-8757-4c4e-8841-37d978cff99a" (UID: "2ee86abf-8757-4c4e-8841-37d978cff99a"). InnerVolumeSpecName "kube-api-access-qx4xc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:29:34 crc kubenswrapper[4578]: I1003 13:29:34.083969 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ee86abf-8757-4c4e-8841-37d978cff99a-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "2ee86abf-8757-4c4e-8841-37d978cff99a" (UID: "2ee86abf-8757-4c4e-8841-37d978cff99a"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:29:34 crc kubenswrapper[4578]: I1003 13:29:34.103075 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ee86abf-8757-4c4e-8841-37d978cff99a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2ee86abf-8757-4c4e-8841-37d978cff99a" (UID: "2ee86abf-8757-4c4e-8841-37d978cff99a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:29:34 crc kubenswrapper[4578]: I1003 13:29:34.107763 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ee86abf-8757-4c4e-8841-37d978cff99a-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "2ee86abf-8757-4c4e-8841-37d978cff99a" (UID: "2ee86abf-8757-4c4e-8841-37d978cff99a"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:29:34 crc kubenswrapper[4578]: I1003 13:29:34.108525 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ee86abf-8757-4c4e-8841-37d978cff99a-inventory" (OuterVolumeSpecName: "inventory") pod "2ee86abf-8757-4c4e-8841-37d978cff99a" (UID: "2ee86abf-8757-4c4e-8841-37d978cff99a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:29:34 crc kubenswrapper[4578]: I1003 13:29:34.178004 4578 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ee86abf-8757-4c4e-8841-37d978cff99a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:29:34 crc kubenswrapper[4578]: I1003 13:29:34.178055 4578 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ee86abf-8757-4c4e-8841-37d978cff99a-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 13:29:34 crc kubenswrapper[4578]: I1003 13:29:34.178067 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qx4xc\" (UniqueName: \"kubernetes.io/projected/2ee86abf-8757-4c4e-8841-37d978cff99a-kube-api-access-qx4xc\") on node \"crc\" DevicePath \"\"" Oct 03 13:29:34 crc kubenswrapper[4578]: I1003 13:29:34.178083 4578 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee86abf-8757-4c4e-8841-37d978cff99a-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:29:34 crc kubenswrapper[4578]: I1003 13:29:34.178094 4578 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/2ee86abf-8757-4c4e-8841-37d978cff99a-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:29:34 crc kubenswrapper[4578]: I1003 13:29:34.540772 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" event={"ID":"2ee86abf-8757-4c4e-8841-37d978cff99a","Type":"ContainerDied","Data":"2faaae5f732f849c8be55a995044f7009ba0e92dd1bb13e72f46c0e25305cb52"} Oct 03 13:29:34 crc kubenswrapper[4578]: I1003 13:29:34.541130 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2faaae5f732f849c8be55a995044f7009ba0e92dd1bb13e72f46c0e25305cb52" Oct 03 13:29:34 crc kubenswrapper[4578]: I1003 13:29:34.540912 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-tnz6b" Oct 03 13:29:36 crc kubenswrapper[4578]: I1003 13:29:36.909804 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:29:36 crc kubenswrapper[4578]: E1003 13:29:36.910512 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.025809 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv"] Oct 03 13:29:42 crc kubenswrapper[4578]: E1003 13:29:42.026776 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee86abf-8757-4c4e-8841-37d978cff99a" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.026798 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee86abf-8757-4c4e-8841-37d978cff99a" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.027028 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ee86abf-8757-4c4e-8841-37d978cff99a" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.027692 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.031215 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.032571 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.033278 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-442mn" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.033331 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.034368 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.039348 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv"] Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.148400 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8856883f-4c07-432c-aadf-399e845876cf-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7p4mv\" (UID: \"8856883f-4c07-432c-aadf-399e845876cf\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.148490 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pft9d\" (UniqueName: \"kubernetes.io/projected/8856883f-4c07-432c-aadf-399e845876cf-kube-api-access-pft9d\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7p4mv\" (UID: \"8856883f-4c07-432c-aadf-399e845876cf\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.148529 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8856883f-4c07-432c-aadf-399e845876cf-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7p4mv\" (UID: \"8856883f-4c07-432c-aadf-399e845876cf\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.148577 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8856883f-4c07-432c-aadf-399e845876cf-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7p4mv\" (UID: \"8856883f-4c07-432c-aadf-399e845876cf\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.148886 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/8856883f-4c07-432c-aadf-399e845876cf-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7p4mv\" (UID: \"8856883f-4c07-432c-aadf-399e845876cf\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.251058 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pft9d\" (UniqueName: \"kubernetes.io/projected/8856883f-4c07-432c-aadf-399e845876cf-kube-api-access-pft9d\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7p4mv\" (UID: \"8856883f-4c07-432c-aadf-399e845876cf\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.251124 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8856883f-4c07-432c-aadf-399e845876cf-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7p4mv\" (UID: \"8856883f-4c07-432c-aadf-399e845876cf\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.251183 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8856883f-4c07-432c-aadf-399e845876cf-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7p4mv\" (UID: \"8856883f-4c07-432c-aadf-399e845876cf\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.251248 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/8856883f-4c07-432c-aadf-399e845876cf-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7p4mv\" (UID: \"8856883f-4c07-432c-aadf-399e845876cf\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.251288 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8856883f-4c07-432c-aadf-399e845876cf-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7p4mv\" (UID: \"8856883f-4c07-432c-aadf-399e845876cf\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.252803 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/8856883f-4c07-432c-aadf-399e845876cf-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7p4mv\" (UID: \"8856883f-4c07-432c-aadf-399e845876cf\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.257494 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8856883f-4c07-432c-aadf-399e845876cf-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7p4mv\" (UID: \"8856883f-4c07-432c-aadf-399e845876cf\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.258952 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8856883f-4c07-432c-aadf-399e845876cf-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7p4mv\" (UID: \"8856883f-4c07-432c-aadf-399e845876cf\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.261119 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8856883f-4c07-432c-aadf-399e845876cf-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7p4mv\" (UID: \"8856883f-4c07-432c-aadf-399e845876cf\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.270401 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pft9d\" (UniqueName: \"kubernetes.io/projected/8856883f-4c07-432c-aadf-399e845876cf-kube-api-access-pft9d\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7p4mv\" (UID: \"8856883f-4c07-432c-aadf-399e845876cf\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.353196 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" Oct 03 13:29:42 crc kubenswrapper[4578]: I1003 13:29:42.905654 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv"] Oct 03 13:29:43 crc kubenswrapper[4578]: I1003 13:29:43.624331 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" event={"ID":"8856883f-4c07-432c-aadf-399e845876cf","Type":"ContainerStarted","Data":"cfe155d5ecf247b4a7e867bc0a648a9cef0d8e99cf260de7fd83b3c555406a02"} Oct 03 13:29:43 crc kubenswrapper[4578]: I1003 13:29:43.624661 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" event={"ID":"8856883f-4c07-432c-aadf-399e845876cf","Type":"ContainerStarted","Data":"0521fa136d640d87701ca996ceb47254152b55dd633b7bffdf21267f5a161f34"} Oct 03 13:29:43 crc kubenswrapper[4578]: I1003 13:29:43.643578 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" podStartSLOduration=1.493413349 podStartE2EDuration="1.64355894s" podCreationTimestamp="2025-10-03 13:29:42 +0000 UTC" firstStartedPulling="2025-10-03 13:29:42.913265784 +0000 UTC m=+2318.711737968" lastFinishedPulling="2025-10-03 13:29:43.063411375 +0000 UTC m=+2318.861883559" observedRunningTime="2025-10-03 13:29:43.640855355 +0000 UTC m=+2319.439327549" watchObservedRunningTime="2025-10-03 13:29:43.64355894 +0000 UTC m=+2319.442031124" Oct 03 13:29:49 crc kubenswrapper[4578]: I1003 13:29:49.908843 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:29:49 crc kubenswrapper[4578]: E1003 13:29:49.909721 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:29:59 crc kubenswrapper[4578]: I1003 13:29:59.865512 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hnhml"] Oct 03 13:29:59 crc kubenswrapper[4578]: I1003 13:29:59.871144 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hnhml" Oct 03 13:29:59 crc kubenswrapper[4578]: I1003 13:29:59.876105 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hnhml"] Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.029128 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87cd531e-9db3-4dd1-af71-32b46328f7dc-utilities\") pod \"redhat-marketplace-hnhml\" (UID: \"87cd531e-9db3-4dd1-af71-32b46328f7dc\") " pod="openshift-marketplace/redhat-marketplace-hnhml" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.029486 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq696\" (UniqueName: \"kubernetes.io/projected/87cd531e-9db3-4dd1-af71-32b46328f7dc-kube-api-access-sq696\") pod \"redhat-marketplace-hnhml\" (UID: \"87cd531e-9db3-4dd1-af71-32b46328f7dc\") " pod="openshift-marketplace/redhat-marketplace-hnhml" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.029773 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87cd531e-9db3-4dd1-af71-32b46328f7dc-catalog-content\") pod \"redhat-marketplace-hnhml\" (UID: \"87cd531e-9db3-4dd1-af71-32b46328f7dc\") " pod="openshift-marketplace/redhat-marketplace-hnhml" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.132687 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87cd531e-9db3-4dd1-af71-32b46328f7dc-catalog-content\") pod \"redhat-marketplace-hnhml\" (UID: \"87cd531e-9db3-4dd1-af71-32b46328f7dc\") " pod="openshift-marketplace/redhat-marketplace-hnhml" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.132896 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87cd531e-9db3-4dd1-af71-32b46328f7dc-utilities\") pod \"redhat-marketplace-hnhml\" (UID: \"87cd531e-9db3-4dd1-af71-32b46328f7dc\") " pod="openshift-marketplace/redhat-marketplace-hnhml" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.133223 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq696\" (UniqueName: \"kubernetes.io/projected/87cd531e-9db3-4dd1-af71-32b46328f7dc-kube-api-access-sq696\") pod \"redhat-marketplace-hnhml\" (UID: \"87cd531e-9db3-4dd1-af71-32b46328f7dc\") " pod="openshift-marketplace/redhat-marketplace-hnhml" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.133292 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87cd531e-9db3-4dd1-af71-32b46328f7dc-catalog-content\") pod \"redhat-marketplace-hnhml\" (UID: \"87cd531e-9db3-4dd1-af71-32b46328f7dc\") " pod="openshift-marketplace/redhat-marketplace-hnhml" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.133417 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87cd531e-9db3-4dd1-af71-32b46328f7dc-utilities\") pod \"redhat-marketplace-hnhml\" (UID: \"87cd531e-9db3-4dd1-af71-32b46328f7dc\") " pod="openshift-marketplace/redhat-marketplace-hnhml" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.155100 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l"] Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.158450 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.168139 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq696\" (UniqueName: \"kubernetes.io/projected/87cd531e-9db3-4dd1-af71-32b46328f7dc-kube-api-access-sq696\") pod \"redhat-marketplace-hnhml\" (UID: \"87cd531e-9db3-4dd1-af71-32b46328f7dc\") " pod="openshift-marketplace/redhat-marketplace-hnhml" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.168794 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.172449 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.175218 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l"] Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.213083 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hnhml" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.338727 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c92e0eba-3ff4-404b-a8ab-064f5ae88a12-secret-volume\") pod \"collect-profiles-29324970-bd42l\" (UID: \"c92e0eba-3ff4-404b-a8ab-064f5ae88a12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.339179 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s26d2\" (UniqueName: \"kubernetes.io/projected/c92e0eba-3ff4-404b-a8ab-064f5ae88a12-kube-api-access-s26d2\") pod \"collect-profiles-29324970-bd42l\" (UID: \"c92e0eba-3ff4-404b-a8ab-064f5ae88a12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.339256 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c92e0eba-3ff4-404b-a8ab-064f5ae88a12-config-volume\") pod \"collect-profiles-29324970-bd42l\" (UID: \"c92e0eba-3ff4-404b-a8ab-064f5ae88a12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.442754 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c92e0eba-3ff4-404b-a8ab-064f5ae88a12-config-volume\") pod \"collect-profiles-29324970-bd42l\" (UID: \"c92e0eba-3ff4-404b-a8ab-064f5ae88a12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.442927 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c92e0eba-3ff4-404b-a8ab-064f5ae88a12-secret-volume\") pod \"collect-profiles-29324970-bd42l\" (UID: \"c92e0eba-3ff4-404b-a8ab-064f5ae88a12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.443096 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s26d2\" (UniqueName: \"kubernetes.io/projected/c92e0eba-3ff4-404b-a8ab-064f5ae88a12-kube-api-access-s26d2\") pod \"collect-profiles-29324970-bd42l\" (UID: \"c92e0eba-3ff4-404b-a8ab-064f5ae88a12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.443908 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c92e0eba-3ff4-404b-a8ab-064f5ae88a12-config-volume\") pod \"collect-profiles-29324970-bd42l\" (UID: \"c92e0eba-3ff4-404b-a8ab-064f5ae88a12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.463216 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s26d2\" (UniqueName: \"kubernetes.io/projected/c92e0eba-3ff4-404b-a8ab-064f5ae88a12-kube-api-access-s26d2\") pod \"collect-profiles-29324970-bd42l\" (UID: \"c92e0eba-3ff4-404b-a8ab-064f5ae88a12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.467857 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c92e0eba-3ff4-404b-a8ab-064f5ae88a12-secret-volume\") pod \"collect-profiles-29324970-bd42l\" (UID: \"c92e0eba-3ff4-404b-a8ab-064f5ae88a12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.541740 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.728809 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hnhml"] Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.770402 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hnhml" event={"ID":"87cd531e-9db3-4dd1-af71-32b46328f7dc","Type":"ContainerStarted","Data":"002ac96782303371031b212aca5bb63198e16f4efd8c617178338a28f2d2d42d"} Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.910462 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:30:00 crc kubenswrapper[4578]: E1003 13:30:00.910768 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:30:00 crc kubenswrapper[4578]: I1003 13:30:00.996193 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l"] Oct 03 13:30:00 crc kubenswrapper[4578]: W1003 13:30:00.999708 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc92e0eba_3ff4_404b_a8ab_064f5ae88a12.slice/crio-ae131a45de85bb27aa59d15956acdf7497d39d63a2ac68f7543e5b98ce604bc1 WatchSource:0}: Error finding container ae131a45de85bb27aa59d15956acdf7497d39d63a2ac68f7543e5b98ce604bc1: Status 404 returned error can't find the container with id ae131a45de85bb27aa59d15956acdf7497d39d63a2ac68f7543e5b98ce604bc1 Oct 03 13:30:01 crc kubenswrapper[4578]: I1003 13:30:01.780901 4578 generic.go:334] "Generic (PLEG): container finished" podID="87cd531e-9db3-4dd1-af71-32b46328f7dc" containerID="b6800d667b3cf45bf22f86a8aff960ed73fb18e7f368d4838f3c392136d1624f" exitCode=0 Oct 03 13:30:01 crc kubenswrapper[4578]: I1003 13:30:01.781002 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hnhml" event={"ID":"87cd531e-9db3-4dd1-af71-32b46328f7dc","Type":"ContainerDied","Data":"b6800d667b3cf45bf22f86a8aff960ed73fb18e7f368d4838f3c392136d1624f"} Oct 03 13:30:01 crc kubenswrapper[4578]: I1003 13:30:01.790247 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l" event={"ID":"c92e0eba-3ff4-404b-a8ab-064f5ae88a12","Type":"ContainerStarted","Data":"5cd906a0cd1fb7de0767611f43fe9ff637558dda5c1e1255bbcb7df80532e249"} Oct 03 13:30:01 crc kubenswrapper[4578]: I1003 13:30:01.790292 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l" event={"ID":"c92e0eba-3ff4-404b-a8ab-064f5ae88a12","Type":"ContainerStarted","Data":"ae131a45de85bb27aa59d15956acdf7497d39d63a2ac68f7543e5b98ce604bc1"} Oct 03 13:30:01 crc kubenswrapper[4578]: I1003 13:30:01.829444 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l" podStartSLOduration=1.8294191180000001 podStartE2EDuration="1.829419118s" podCreationTimestamp="2025-10-03 13:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 13:30:01.820038934 +0000 UTC m=+2337.618511128" watchObservedRunningTime="2025-10-03 13:30:01.829419118 +0000 UTC m=+2337.627891302" Oct 03 13:30:02 crc kubenswrapper[4578]: I1003 13:30:02.800251 4578 generic.go:334] "Generic (PLEG): container finished" podID="c92e0eba-3ff4-404b-a8ab-064f5ae88a12" containerID="5cd906a0cd1fb7de0767611f43fe9ff637558dda5c1e1255bbcb7df80532e249" exitCode=0 Oct 03 13:30:02 crc kubenswrapper[4578]: I1003 13:30:02.800301 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l" event={"ID":"c92e0eba-3ff4-404b-a8ab-064f5ae88a12","Type":"ContainerDied","Data":"5cd906a0cd1fb7de0767611f43fe9ff637558dda5c1e1255bbcb7df80532e249"} Oct 03 13:30:03 crc kubenswrapper[4578]: I1003 13:30:03.809967 4578 generic.go:334] "Generic (PLEG): container finished" podID="87cd531e-9db3-4dd1-af71-32b46328f7dc" containerID="53c82dd1f47c0f80a36aa2769bbe395892778a7fd44a0b82f338a538abc74e04" exitCode=0 Oct 03 13:30:03 crc kubenswrapper[4578]: I1003 13:30:03.810197 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hnhml" event={"ID":"87cd531e-9db3-4dd1-af71-32b46328f7dc","Type":"ContainerDied","Data":"53c82dd1f47c0f80a36aa2769bbe395892778a7fd44a0b82f338a538abc74e04"} Oct 03 13:30:04 crc kubenswrapper[4578]: I1003 13:30:04.154179 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l" Oct 03 13:30:04 crc kubenswrapper[4578]: I1003 13:30:04.324349 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c92e0eba-3ff4-404b-a8ab-064f5ae88a12-config-volume\") pod \"c92e0eba-3ff4-404b-a8ab-064f5ae88a12\" (UID: \"c92e0eba-3ff4-404b-a8ab-064f5ae88a12\") " Oct 03 13:30:04 crc kubenswrapper[4578]: I1003 13:30:04.324541 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s26d2\" (UniqueName: \"kubernetes.io/projected/c92e0eba-3ff4-404b-a8ab-064f5ae88a12-kube-api-access-s26d2\") pod \"c92e0eba-3ff4-404b-a8ab-064f5ae88a12\" (UID: \"c92e0eba-3ff4-404b-a8ab-064f5ae88a12\") " Oct 03 13:30:04 crc kubenswrapper[4578]: I1003 13:30:04.324800 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c92e0eba-3ff4-404b-a8ab-064f5ae88a12-secret-volume\") pod \"c92e0eba-3ff4-404b-a8ab-064f5ae88a12\" (UID: \"c92e0eba-3ff4-404b-a8ab-064f5ae88a12\") " Oct 03 13:30:04 crc kubenswrapper[4578]: I1003 13:30:04.325147 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c92e0eba-3ff4-404b-a8ab-064f5ae88a12-config-volume" (OuterVolumeSpecName: "config-volume") pod "c92e0eba-3ff4-404b-a8ab-064f5ae88a12" (UID: "c92e0eba-3ff4-404b-a8ab-064f5ae88a12"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:04 crc kubenswrapper[4578]: I1003 13:30:04.325262 4578 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c92e0eba-3ff4-404b-a8ab-064f5ae88a12-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:04 crc kubenswrapper[4578]: I1003 13:30:04.330957 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c92e0eba-3ff4-404b-a8ab-064f5ae88a12-kube-api-access-s26d2" (OuterVolumeSpecName: "kube-api-access-s26d2") pod "c92e0eba-3ff4-404b-a8ab-064f5ae88a12" (UID: "c92e0eba-3ff4-404b-a8ab-064f5ae88a12"). InnerVolumeSpecName "kube-api-access-s26d2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:04 crc kubenswrapper[4578]: I1003 13:30:04.331093 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c92e0eba-3ff4-404b-a8ab-064f5ae88a12-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c92e0eba-3ff4-404b-a8ab-064f5ae88a12" (UID: "c92e0eba-3ff4-404b-a8ab-064f5ae88a12"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:04 crc kubenswrapper[4578]: I1003 13:30:04.427369 4578 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c92e0eba-3ff4-404b-a8ab-064f5ae88a12-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:04 crc kubenswrapper[4578]: I1003 13:30:04.427987 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s26d2\" (UniqueName: \"kubernetes.io/projected/c92e0eba-3ff4-404b-a8ab-064f5ae88a12-kube-api-access-s26d2\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:04 crc kubenswrapper[4578]: I1003 13:30:04.820116 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hnhml" event={"ID":"87cd531e-9db3-4dd1-af71-32b46328f7dc","Type":"ContainerStarted","Data":"71222e15d11393d98389944f5f0acd58ce7a6d3b963b4688d837e900a1a3e468"} Oct 03 13:30:04 crc kubenswrapper[4578]: I1003 13:30:04.822592 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l" event={"ID":"c92e0eba-3ff4-404b-a8ab-064f5ae88a12","Type":"ContainerDied","Data":"ae131a45de85bb27aa59d15956acdf7497d39d63a2ac68f7543e5b98ce604bc1"} Oct 03 13:30:04 crc kubenswrapper[4578]: I1003 13:30:04.822621 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae131a45de85bb27aa59d15956acdf7497d39d63a2ac68f7543e5b98ce604bc1" Oct 03 13:30:04 crc kubenswrapper[4578]: I1003 13:30:04.822683 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l" Oct 03 13:30:04 crc kubenswrapper[4578]: I1003 13:30:04.848248 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hnhml" podStartSLOduration=3.375186894 podStartE2EDuration="5.848231848s" podCreationTimestamp="2025-10-03 13:29:59 +0000 UTC" firstStartedPulling="2025-10-03 13:30:01.787823213 +0000 UTC m=+2337.586295397" lastFinishedPulling="2025-10-03 13:30:04.260868167 +0000 UTC m=+2340.059340351" observedRunningTime="2025-10-03 13:30:04.841213218 +0000 UTC m=+2340.639685402" watchObservedRunningTime="2025-10-03 13:30:04.848231848 +0000 UTC m=+2340.646704032" Oct 03 13:30:05 crc kubenswrapper[4578]: I1003 13:30:05.227250 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4"] Oct 03 13:30:05 crc kubenswrapper[4578]: I1003 13:30:05.235993 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324925-64fg4"] Oct 03 13:30:06 crc kubenswrapper[4578]: I1003 13:30:06.922269 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdfd4313-d05e-4cbd-8520-3f929675c662" path="/var/lib/kubelet/pods/fdfd4313-d05e-4cbd-8520-3f929675c662/volumes" Oct 03 13:30:09 crc kubenswrapper[4578]: I1003 13:30:09.444792 4578 scope.go:117] "RemoveContainer" containerID="05ee2bd4e95415aa471a093fa842b223d3543a457c885f2d0816faaf04cf5fd6" Oct 03 13:30:10 crc kubenswrapper[4578]: I1003 13:30:10.214052 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hnhml" Oct 03 13:30:10 crc kubenswrapper[4578]: I1003 13:30:10.214345 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hnhml" Oct 03 13:30:10 crc kubenswrapper[4578]: I1003 13:30:10.269186 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hnhml" Oct 03 13:30:10 crc kubenswrapper[4578]: I1003 13:30:10.918199 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hnhml" Oct 03 13:30:10 crc kubenswrapper[4578]: I1003 13:30:10.976160 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hnhml"] Oct 03 13:30:11 crc kubenswrapper[4578]: I1003 13:30:11.909939 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:30:11 crc kubenswrapper[4578]: E1003 13:30:11.911001 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:30:12 crc kubenswrapper[4578]: I1003 13:30:12.889370 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hnhml" podUID="87cd531e-9db3-4dd1-af71-32b46328f7dc" containerName="registry-server" containerID="cri-o://71222e15d11393d98389944f5f0acd58ce7a6d3b963b4688d837e900a1a3e468" gracePeriod=2 Oct 03 13:30:13 crc kubenswrapper[4578]: I1003 13:30:13.855504 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hnhml" Oct 03 13:30:13 crc kubenswrapper[4578]: I1003 13:30:13.912043 4578 generic.go:334] "Generic (PLEG): container finished" podID="87cd531e-9db3-4dd1-af71-32b46328f7dc" containerID="71222e15d11393d98389944f5f0acd58ce7a6d3b963b4688d837e900a1a3e468" exitCode=0 Oct 03 13:30:13 crc kubenswrapper[4578]: I1003 13:30:13.912121 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hnhml" event={"ID":"87cd531e-9db3-4dd1-af71-32b46328f7dc","Type":"ContainerDied","Data":"71222e15d11393d98389944f5f0acd58ce7a6d3b963b4688d837e900a1a3e468"} Oct 03 13:30:13 crc kubenswrapper[4578]: I1003 13:30:13.912828 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hnhml" event={"ID":"87cd531e-9db3-4dd1-af71-32b46328f7dc","Type":"ContainerDied","Data":"002ac96782303371031b212aca5bb63198e16f4efd8c617178338a28f2d2d42d"} Oct 03 13:30:13 crc kubenswrapper[4578]: I1003 13:30:13.912165 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hnhml" Oct 03 13:30:13 crc kubenswrapper[4578]: I1003 13:30:13.912862 4578 scope.go:117] "RemoveContainer" containerID="71222e15d11393d98389944f5f0acd58ce7a6d3b963b4688d837e900a1a3e468" Oct 03 13:30:13 crc kubenswrapper[4578]: I1003 13:30:13.917242 4578 generic.go:334] "Generic (PLEG): container finished" podID="8856883f-4c07-432c-aadf-399e845876cf" containerID="cfe155d5ecf247b4a7e867bc0a648a9cef0d8e99cf260de7fd83b3c555406a02" exitCode=2 Oct 03 13:30:13 crc kubenswrapper[4578]: I1003 13:30:13.917279 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" event={"ID":"8856883f-4c07-432c-aadf-399e845876cf","Type":"ContainerDied","Data":"cfe155d5ecf247b4a7e867bc0a648a9cef0d8e99cf260de7fd83b3c555406a02"} Oct 03 13:30:13 crc kubenswrapper[4578]: I1003 13:30:13.954308 4578 scope.go:117] "RemoveContainer" containerID="53c82dd1f47c0f80a36aa2769bbe395892778a7fd44a0b82f338a538abc74e04" Oct 03 13:30:13 crc kubenswrapper[4578]: I1003 13:30:13.989918 4578 scope.go:117] "RemoveContainer" containerID="b6800d667b3cf45bf22f86a8aff960ed73fb18e7f368d4838f3c392136d1624f" Oct 03 13:30:14 crc kubenswrapper[4578]: I1003 13:30:14.000910 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87cd531e-9db3-4dd1-af71-32b46328f7dc-utilities\") pod \"87cd531e-9db3-4dd1-af71-32b46328f7dc\" (UID: \"87cd531e-9db3-4dd1-af71-32b46328f7dc\") " Oct 03 13:30:14 crc kubenswrapper[4578]: I1003 13:30:14.001049 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sq696\" (UniqueName: \"kubernetes.io/projected/87cd531e-9db3-4dd1-af71-32b46328f7dc-kube-api-access-sq696\") pod \"87cd531e-9db3-4dd1-af71-32b46328f7dc\" (UID: \"87cd531e-9db3-4dd1-af71-32b46328f7dc\") " Oct 03 13:30:14 crc kubenswrapper[4578]: I1003 13:30:14.001111 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87cd531e-9db3-4dd1-af71-32b46328f7dc-catalog-content\") pod \"87cd531e-9db3-4dd1-af71-32b46328f7dc\" (UID: \"87cd531e-9db3-4dd1-af71-32b46328f7dc\") " Oct 03 13:30:14 crc kubenswrapper[4578]: I1003 13:30:14.001913 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87cd531e-9db3-4dd1-af71-32b46328f7dc-utilities" (OuterVolumeSpecName: "utilities") pod "87cd531e-9db3-4dd1-af71-32b46328f7dc" (UID: "87cd531e-9db3-4dd1-af71-32b46328f7dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:30:14 crc kubenswrapper[4578]: I1003 13:30:14.009153 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cd531e-9db3-4dd1-af71-32b46328f7dc-kube-api-access-sq696" (OuterVolumeSpecName: "kube-api-access-sq696") pod "87cd531e-9db3-4dd1-af71-32b46328f7dc" (UID: "87cd531e-9db3-4dd1-af71-32b46328f7dc"). InnerVolumeSpecName "kube-api-access-sq696". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:14 crc kubenswrapper[4578]: I1003 13:30:14.016194 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87cd531e-9db3-4dd1-af71-32b46328f7dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "87cd531e-9db3-4dd1-af71-32b46328f7dc" (UID: "87cd531e-9db3-4dd1-af71-32b46328f7dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:30:14 crc kubenswrapper[4578]: I1003 13:30:14.070516 4578 scope.go:117] "RemoveContainer" containerID="71222e15d11393d98389944f5f0acd58ce7a6d3b963b4688d837e900a1a3e468" Oct 03 13:30:14 crc kubenswrapper[4578]: E1003 13:30:14.071033 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71222e15d11393d98389944f5f0acd58ce7a6d3b963b4688d837e900a1a3e468\": container with ID starting with 71222e15d11393d98389944f5f0acd58ce7a6d3b963b4688d837e900a1a3e468 not found: ID does not exist" containerID="71222e15d11393d98389944f5f0acd58ce7a6d3b963b4688d837e900a1a3e468" Oct 03 13:30:14 crc kubenswrapper[4578]: I1003 13:30:14.071077 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71222e15d11393d98389944f5f0acd58ce7a6d3b963b4688d837e900a1a3e468"} err="failed to get container status \"71222e15d11393d98389944f5f0acd58ce7a6d3b963b4688d837e900a1a3e468\": rpc error: code = NotFound desc = could not find container \"71222e15d11393d98389944f5f0acd58ce7a6d3b963b4688d837e900a1a3e468\": container with ID starting with 71222e15d11393d98389944f5f0acd58ce7a6d3b963b4688d837e900a1a3e468 not found: ID does not exist" Oct 03 13:30:14 crc kubenswrapper[4578]: I1003 13:30:14.071103 4578 scope.go:117] "RemoveContainer" containerID="53c82dd1f47c0f80a36aa2769bbe395892778a7fd44a0b82f338a538abc74e04" Oct 03 13:30:14 crc kubenswrapper[4578]: E1003 13:30:14.071543 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53c82dd1f47c0f80a36aa2769bbe395892778a7fd44a0b82f338a538abc74e04\": container with ID starting with 53c82dd1f47c0f80a36aa2769bbe395892778a7fd44a0b82f338a538abc74e04 not found: ID does not exist" containerID="53c82dd1f47c0f80a36aa2769bbe395892778a7fd44a0b82f338a538abc74e04" Oct 03 13:30:14 crc kubenswrapper[4578]: I1003 13:30:14.071614 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53c82dd1f47c0f80a36aa2769bbe395892778a7fd44a0b82f338a538abc74e04"} err="failed to get container status \"53c82dd1f47c0f80a36aa2769bbe395892778a7fd44a0b82f338a538abc74e04\": rpc error: code = NotFound desc = could not find container \"53c82dd1f47c0f80a36aa2769bbe395892778a7fd44a0b82f338a538abc74e04\": container with ID starting with 53c82dd1f47c0f80a36aa2769bbe395892778a7fd44a0b82f338a538abc74e04 not found: ID does not exist" Oct 03 13:30:14 crc kubenswrapper[4578]: I1003 13:30:14.071669 4578 scope.go:117] "RemoveContainer" containerID="b6800d667b3cf45bf22f86a8aff960ed73fb18e7f368d4838f3c392136d1624f" Oct 03 13:30:14 crc kubenswrapper[4578]: E1003 13:30:14.072285 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6800d667b3cf45bf22f86a8aff960ed73fb18e7f368d4838f3c392136d1624f\": container with ID starting with b6800d667b3cf45bf22f86a8aff960ed73fb18e7f368d4838f3c392136d1624f not found: ID does not exist" containerID="b6800d667b3cf45bf22f86a8aff960ed73fb18e7f368d4838f3c392136d1624f" Oct 03 13:30:14 crc kubenswrapper[4578]: I1003 13:30:14.072315 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6800d667b3cf45bf22f86a8aff960ed73fb18e7f368d4838f3c392136d1624f"} err="failed to get container status \"b6800d667b3cf45bf22f86a8aff960ed73fb18e7f368d4838f3c392136d1624f\": rpc error: code = NotFound desc = could not find container \"b6800d667b3cf45bf22f86a8aff960ed73fb18e7f368d4838f3c392136d1624f\": container with ID starting with b6800d667b3cf45bf22f86a8aff960ed73fb18e7f368d4838f3c392136d1624f not found: ID does not exist" Oct 03 13:30:14 crc kubenswrapper[4578]: I1003 13:30:14.103111 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87cd531e-9db3-4dd1-af71-32b46328f7dc-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:14 crc kubenswrapper[4578]: I1003 13:30:14.103143 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sq696\" (UniqueName: \"kubernetes.io/projected/87cd531e-9db3-4dd1-af71-32b46328f7dc-kube-api-access-sq696\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:14 crc kubenswrapper[4578]: I1003 13:30:14.103154 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87cd531e-9db3-4dd1-af71-32b46328f7dc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:14 crc kubenswrapper[4578]: I1003 13:30:14.244088 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hnhml"] Oct 03 13:30:14 crc kubenswrapper[4578]: I1003 13:30:14.253401 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hnhml"] Oct 03 13:30:14 crc kubenswrapper[4578]: I1003 13:30:14.921910 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cd531e-9db3-4dd1-af71-32b46328f7dc" path="/var/lib/kubelet/pods/87cd531e-9db3-4dd1-af71-32b46328f7dc/volumes" Oct 03 13:30:15 crc kubenswrapper[4578]: I1003 13:30:15.307889 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" Oct 03 13:30:15 crc kubenswrapper[4578]: I1003 13:30:15.428486 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/8856883f-4c07-432c-aadf-399e845876cf-ovncontroller-config-0\") pod \"8856883f-4c07-432c-aadf-399e845876cf\" (UID: \"8856883f-4c07-432c-aadf-399e845876cf\") " Oct 03 13:30:15 crc kubenswrapper[4578]: I1003 13:30:15.428582 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8856883f-4c07-432c-aadf-399e845876cf-ssh-key\") pod \"8856883f-4c07-432c-aadf-399e845876cf\" (UID: \"8856883f-4c07-432c-aadf-399e845876cf\") " Oct 03 13:30:15 crc kubenswrapper[4578]: I1003 13:30:15.428697 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8856883f-4c07-432c-aadf-399e845876cf-inventory\") pod \"8856883f-4c07-432c-aadf-399e845876cf\" (UID: \"8856883f-4c07-432c-aadf-399e845876cf\") " Oct 03 13:30:15 crc kubenswrapper[4578]: I1003 13:30:15.428778 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8856883f-4c07-432c-aadf-399e845876cf-ovn-combined-ca-bundle\") pod \"8856883f-4c07-432c-aadf-399e845876cf\" (UID: \"8856883f-4c07-432c-aadf-399e845876cf\") " Oct 03 13:30:15 crc kubenswrapper[4578]: I1003 13:30:15.428940 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pft9d\" (UniqueName: \"kubernetes.io/projected/8856883f-4c07-432c-aadf-399e845876cf-kube-api-access-pft9d\") pod \"8856883f-4c07-432c-aadf-399e845876cf\" (UID: \"8856883f-4c07-432c-aadf-399e845876cf\") " Oct 03 13:30:15 crc kubenswrapper[4578]: I1003 13:30:15.433775 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8856883f-4c07-432c-aadf-399e845876cf-kube-api-access-pft9d" (OuterVolumeSpecName: "kube-api-access-pft9d") pod "8856883f-4c07-432c-aadf-399e845876cf" (UID: "8856883f-4c07-432c-aadf-399e845876cf"). InnerVolumeSpecName "kube-api-access-pft9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:30:15 crc kubenswrapper[4578]: I1003 13:30:15.445936 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8856883f-4c07-432c-aadf-399e845876cf-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "8856883f-4c07-432c-aadf-399e845876cf" (UID: "8856883f-4c07-432c-aadf-399e845876cf"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:15 crc kubenswrapper[4578]: I1003 13:30:15.459610 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8856883f-4c07-432c-aadf-399e845876cf-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "8856883f-4c07-432c-aadf-399e845876cf" (UID: "8856883f-4c07-432c-aadf-399e845876cf"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:30:15 crc kubenswrapper[4578]: I1003 13:30:15.460696 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8856883f-4c07-432c-aadf-399e845876cf-inventory" (OuterVolumeSpecName: "inventory") pod "8856883f-4c07-432c-aadf-399e845876cf" (UID: "8856883f-4c07-432c-aadf-399e845876cf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:15 crc kubenswrapper[4578]: I1003 13:30:15.461164 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8856883f-4c07-432c-aadf-399e845876cf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8856883f-4c07-432c-aadf-399e845876cf" (UID: "8856883f-4c07-432c-aadf-399e845876cf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:30:15 crc kubenswrapper[4578]: I1003 13:30:15.531735 4578 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8856883f-4c07-432c-aadf-399e845876cf-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:15 crc kubenswrapper[4578]: I1003 13:30:15.531782 4578 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8856883f-4c07-432c-aadf-399e845876cf-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:15 crc kubenswrapper[4578]: I1003 13:30:15.531799 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pft9d\" (UniqueName: \"kubernetes.io/projected/8856883f-4c07-432c-aadf-399e845876cf-kube-api-access-pft9d\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:15 crc kubenswrapper[4578]: I1003 13:30:15.531810 4578 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/8856883f-4c07-432c-aadf-399e845876cf-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:15 crc kubenswrapper[4578]: I1003 13:30:15.531852 4578 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8856883f-4c07-432c-aadf-399e845876cf-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:30:15 crc kubenswrapper[4578]: I1003 13:30:15.935168 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" Oct 03 13:30:15 crc kubenswrapper[4578]: I1003 13:30:15.935185 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7p4mv" event={"ID":"8856883f-4c07-432c-aadf-399e845876cf","Type":"ContainerDied","Data":"0521fa136d640d87701ca996ceb47254152b55dd633b7bffdf21267f5a161f34"} Oct 03 13:30:15 crc kubenswrapper[4578]: I1003 13:30:15.935237 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0521fa136d640d87701ca996ceb47254152b55dd633b7bffdf21267f5a161f34" Oct 03 13:30:24 crc kubenswrapper[4578]: I1003 13:30:24.916108 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:30:24 crc kubenswrapper[4578]: E1003 13:30:24.916999 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.036992 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl"] Oct 03 13:30:33 crc kubenswrapper[4578]: E1003 13:30:33.037816 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87cd531e-9db3-4dd1-af71-32b46328f7dc" containerName="registry-server" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.037828 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="87cd531e-9db3-4dd1-af71-32b46328f7dc" containerName="registry-server" Oct 03 13:30:33 crc kubenswrapper[4578]: E1003 13:30:33.037850 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87cd531e-9db3-4dd1-af71-32b46328f7dc" containerName="extract-utilities" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.037856 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="87cd531e-9db3-4dd1-af71-32b46328f7dc" containerName="extract-utilities" Oct 03 13:30:33 crc kubenswrapper[4578]: E1003 13:30:33.037876 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87cd531e-9db3-4dd1-af71-32b46328f7dc" containerName="extract-content" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.037883 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="87cd531e-9db3-4dd1-af71-32b46328f7dc" containerName="extract-content" Oct 03 13:30:33 crc kubenswrapper[4578]: E1003 13:30:33.037897 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c92e0eba-3ff4-404b-a8ab-064f5ae88a12" containerName="collect-profiles" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.037903 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="c92e0eba-3ff4-404b-a8ab-064f5ae88a12" containerName="collect-profiles" Oct 03 13:30:33 crc kubenswrapper[4578]: E1003 13:30:33.037942 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8856883f-4c07-432c-aadf-399e845876cf" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.037950 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="8856883f-4c07-432c-aadf-399e845876cf" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.038145 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="87cd531e-9db3-4dd1-af71-32b46328f7dc" containerName="registry-server" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.038162 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="c92e0eba-3ff4-404b-a8ab-064f5ae88a12" containerName="collect-profiles" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.038174 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="8856883f-4c07-432c-aadf-399e845876cf" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.038758 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.040865 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.041102 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-442mn" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.041261 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.041430 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.043178 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.057806 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl"] Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.169872 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3a6aa5b7-9383-4465-b597-f7330e091dcc-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dbcwl\" (UID: \"3a6aa5b7-9383-4465-b597-f7330e091dcc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.170104 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a6aa5b7-9383-4465-b597-f7330e091dcc-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dbcwl\" (UID: \"3a6aa5b7-9383-4465-b597-f7330e091dcc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.170229 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2frh\" (UniqueName: \"kubernetes.io/projected/3a6aa5b7-9383-4465-b597-f7330e091dcc-kube-api-access-m2frh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dbcwl\" (UID: \"3a6aa5b7-9383-4465-b597-f7330e091dcc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.170336 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a6aa5b7-9383-4465-b597-f7330e091dcc-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dbcwl\" (UID: \"3a6aa5b7-9383-4465-b597-f7330e091dcc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.170443 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a6aa5b7-9383-4465-b597-f7330e091dcc-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dbcwl\" (UID: \"3a6aa5b7-9383-4465-b597-f7330e091dcc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.272151 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a6aa5b7-9383-4465-b597-f7330e091dcc-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dbcwl\" (UID: \"3a6aa5b7-9383-4465-b597-f7330e091dcc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.272217 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2frh\" (UniqueName: \"kubernetes.io/projected/3a6aa5b7-9383-4465-b597-f7330e091dcc-kube-api-access-m2frh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dbcwl\" (UID: \"3a6aa5b7-9383-4465-b597-f7330e091dcc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.272253 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a6aa5b7-9383-4465-b597-f7330e091dcc-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dbcwl\" (UID: \"3a6aa5b7-9383-4465-b597-f7330e091dcc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.272301 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a6aa5b7-9383-4465-b597-f7330e091dcc-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dbcwl\" (UID: \"3a6aa5b7-9383-4465-b597-f7330e091dcc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.272427 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3a6aa5b7-9383-4465-b597-f7330e091dcc-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dbcwl\" (UID: \"3a6aa5b7-9383-4465-b597-f7330e091dcc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.273324 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3a6aa5b7-9383-4465-b597-f7330e091dcc-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dbcwl\" (UID: \"3a6aa5b7-9383-4465-b597-f7330e091dcc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.277556 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a6aa5b7-9383-4465-b597-f7330e091dcc-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dbcwl\" (UID: \"3a6aa5b7-9383-4465-b597-f7330e091dcc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.277651 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a6aa5b7-9383-4465-b597-f7330e091dcc-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dbcwl\" (UID: \"3a6aa5b7-9383-4465-b597-f7330e091dcc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.286650 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a6aa5b7-9383-4465-b597-f7330e091dcc-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dbcwl\" (UID: \"3a6aa5b7-9383-4465-b597-f7330e091dcc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.289964 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2frh\" (UniqueName: \"kubernetes.io/projected/3a6aa5b7-9383-4465-b597-f7330e091dcc-kube-api-access-m2frh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dbcwl\" (UID: \"3a6aa5b7-9383-4465-b597-f7330e091dcc\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.377744 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" Oct 03 13:30:33 crc kubenswrapper[4578]: I1003 13:30:33.896741 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl"] Oct 03 13:30:34 crc kubenswrapper[4578]: I1003 13:30:34.104460 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" event={"ID":"3a6aa5b7-9383-4465-b597-f7330e091dcc","Type":"ContainerStarted","Data":"9ffe86d3f52518a19da0140e9b1279f65c6e22dbdcfe208ac54737cbd9f616ee"} Oct 03 13:30:35 crc kubenswrapper[4578]: I1003 13:30:35.116389 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" event={"ID":"3a6aa5b7-9383-4465-b597-f7330e091dcc","Type":"ContainerStarted","Data":"dcf0625ce362f17974e40740a5015c17742d94c90251a9c1caaae5ece2c15be3"} Oct 03 13:30:35 crc kubenswrapper[4578]: I1003 13:30:35.154607 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" podStartSLOduration=1.847787085 podStartE2EDuration="2.154580822s" podCreationTimestamp="2025-10-03 13:30:33 +0000 UTC" firstStartedPulling="2025-10-03 13:30:33.900718277 +0000 UTC m=+2369.699190461" lastFinishedPulling="2025-10-03 13:30:34.207512014 +0000 UTC m=+2370.005984198" observedRunningTime="2025-10-03 13:30:35.135262677 +0000 UTC m=+2370.933734871" watchObservedRunningTime="2025-10-03 13:30:35.154580822 +0000 UTC m=+2370.953053006" Oct 03 13:30:36 crc kubenswrapper[4578]: I1003 13:30:36.909855 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:30:36 crc kubenswrapper[4578]: E1003 13:30:36.910477 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:30:48 crc kubenswrapper[4578]: I1003 13:30:48.909577 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:30:49 crc kubenswrapper[4578]: E1003 13:30:48.910313 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:31:01 crc kubenswrapper[4578]: I1003 13:31:01.909242 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:31:01 crc kubenswrapper[4578]: E1003 13:31:01.909949 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:31:05 crc kubenswrapper[4578]: I1003 13:31:05.373269 4578 generic.go:334] "Generic (PLEG): container finished" podID="3a6aa5b7-9383-4465-b597-f7330e091dcc" containerID="dcf0625ce362f17974e40740a5015c17742d94c90251a9c1caaae5ece2c15be3" exitCode=2 Oct 03 13:31:05 crc kubenswrapper[4578]: I1003 13:31:05.373383 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" event={"ID":"3a6aa5b7-9383-4465-b597-f7330e091dcc","Type":"ContainerDied","Data":"dcf0625ce362f17974e40740a5015c17742d94c90251a9c1caaae5ece2c15be3"} Oct 03 13:31:06 crc kubenswrapper[4578]: I1003 13:31:06.754671 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" Oct 03 13:31:06 crc kubenswrapper[4578]: I1003 13:31:06.797002 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a6aa5b7-9383-4465-b597-f7330e091dcc-ovn-combined-ca-bundle\") pod \"3a6aa5b7-9383-4465-b597-f7330e091dcc\" (UID: \"3a6aa5b7-9383-4465-b597-f7330e091dcc\") " Oct 03 13:31:06 crc kubenswrapper[4578]: I1003 13:31:06.797109 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a6aa5b7-9383-4465-b597-f7330e091dcc-inventory\") pod \"3a6aa5b7-9383-4465-b597-f7330e091dcc\" (UID: \"3a6aa5b7-9383-4465-b597-f7330e091dcc\") " Oct 03 13:31:06 crc kubenswrapper[4578]: I1003 13:31:06.797167 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3a6aa5b7-9383-4465-b597-f7330e091dcc-ovncontroller-config-0\") pod \"3a6aa5b7-9383-4465-b597-f7330e091dcc\" (UID: \"3a6aa5b7-9383-4465-b597-f7330e091dcc\") " Oct 03 13:31:06 crc kubenswrapper[4578]: I1003 13:31:06.797199 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2frh\" (UniqueName: \"kubernetes.io/projected/3a6aa5b7-9383-4465-b597-f7330e091dcc-kube-api-access-m2frh\") pod \"3a6aa5b7-9383-4465-b597-f7330e091dcc\" (UID: \"3a6aa5b7-9383-4465-b597-f7330e091dcc\") " Oct 03 13:31:06 crc kubenswrapper[4578]: I1003 13:31:06.797328 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a6aa5b7-9383-4465-b597-f7330e091dcc-ssh-key\") pod \"3a6aa5b7-9383-4465-b597-f7330e091dcc\" (UID: \"3a6aa5b7-9383-4465-b597-f7330e091dcc\") " Oct 03 13:31:06 crc kubenswrapper[4578]: I1003 13:31:06.803497 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a6aa5b7-9383-4465-b597-f7330e091dcc-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "3a6aa5b7-9383-4465-b597-f7330e091dcc" (UID: "3a6aa5b7-9383-4465-b597-f7330e091dcc"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:31:06 crc kubenswrapper[4578]: I1003 13:31:06.808564 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a6aa5b7-9383-4465-b597-f7330e091dcc-kube-api-access-m2frh" (OuterVolumeSpecName: "kube-api-access-m2frh") pod "3a6aa5b7-9383-4465-b597-f7330e091dcc" (UID: "3a6aa5b7-9383-4465-b597-f7330e091dcc"). InnerVolumeSpecName "kube-api-access-m2frh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:31:06 crc kubenswrapper[4578]: I1003 13:31:06.825271 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a6aa5b7-9383-4465-b597-f7330e091dcc-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "3a6aa5b7-9383-4465-b597-f7330e091dcc" (UID: "3a6aa5b7-9383-4465-b597-f7330e091dcc"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:31:06 crc kubenswrapper[4578]: I1003 13:31:06.827162 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a6aa5b7-9383-4465-b597-f7330e091dcc-inventory" (OuterVolumeSpecName: "inventory") pod "3a6aa5b7-9383-4465-b597-f7330e091dcc" (UID: "3a6aa5b7-9383-4465-b597-f7330e091dcc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:31:06 crc kubenswrapper[4578]: I1003 13:31:06.827842 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a6aa5b7-9383-4465-b597-f7330e091dcc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3a6aa5b7-9383-4465-b597-f7330e091dcc" (UID: "3a6aa5b7-9383-4465-b597-f7330e091dcc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:31:06 crc kubenswrapper[4578]: I1003 13:31:06.900657 4578 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a6aa5b7-9383-4465-b597-f7330e091dcc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:31:06 crc kubenswrapper[4578]: I1003 13:31:06.900983 4578 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a6aa5b7-9383-4465-b597-f7330e091dcc-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:31:06 crc kubenswrapper[4578]: I1003 13:31:06.901056 4578 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a6aa5b7-9383-4465-b597-f7330e091dcc-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 13:31:06 crc kubenswrapper[4578]: I1003 13:31:06.901112 4578 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3a6aa5b7-9383-4465-b597-f7330e091dcc-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:31:06 crc kubenswrapper[4578]: I1003 13:31:06.901177 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2frh\" (UniqueName: \"kubernetes.io/projected/3a6aa5b7-9383-4465-b597-f7330e091dcc-kube-api-access-m2frh\") on node \"crc\" DevicePath \"\"" Oct 03 13:31:07 crc kubenswrapper[4578]: I1003 13:31:07.392766 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" event={"ID":"3a6aa5b7-9383-4465-b597-f7330e091dcc","Type":"ContainerDied","Data":"9ffe86d3f52518a19da0140e9b1279f65c6e22dbdcfe208ac54737cbd9f616ee"} Oct 03 13:31:07 crc kubenswrapper[4578]: I1003 13:31:07.393131 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ffe86d3f52518a19da0140e9b1279f65c6e22dbdcfe208ac54737cbd9f616ee" Oct 03 13:31:07 crc kubenswrapper[4578]: I1003 13:31:07.392891 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dbcwl" Oct 03 13:31:12 crc kubenswrapper[4578]: I1003 13:31:12.909259 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:31:12 crc kubenswrapper[4578]: E1003 13:31:12.911510 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:31:27 crc kubenswrapper[4578]: I1003 13:31:27.909666 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:31:27 crc kubenswrapper[4578]: E1003 13:31:27.910417 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:31:38 crc kubenswrapper[4578]: I1003 13:31:38.909598 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:31:38 crc kubenswrapper[4578]: E1003 13:31:38.910460 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.033552 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw"] Oct 03 13:31:45 crc kubenswrapper[4578]: E1003 13:31:45.034660 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a6aa5b7-9383-4465-b597-f7330e091dcc" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.034678 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a6aa5b7-9383-4465-b597-f7330e091dcc" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.034939 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a6aa5b7-9383-4465-b597-f7330e091dcc" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.035706 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.037795 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.037937 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.038034 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.038450 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-442mn" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.038734 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.055830 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw"] Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.120589 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/084860df-b3eb-43f8-aa12-2b7feea78f6a-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dqxfw\" (UID: \"084860df-b3eb-43f8-aa12-2b7feea78f6a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.120859 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/084860df-b3eb-43f8-aa12-2b7feea78f6a-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dqxfw\" (UID: \"084860df-b3eb-43f8-aa12-2b7feea78f6a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.120996 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/084860df-b3eb-43f8-aa12-2b7feea78f6a-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dqxfw\" (UID: \"084860df-b3eb-43f8-aa12-2b7feea78f6a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.121113 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87gvp\" (UniqueName: \"kubernetes.io/projected/084860df-b3eb-43f8-aa12-2b7feea78f6a-kube-api-access-87gvp\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dqxfw\" (UID: \"084860df-b3eb-43f8-aa12-2b7feea78f6a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.121244 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/084860df-b3eb-43f8-aa12-2b7feea78f6a-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dqxfw\" (UID: \"084860df-b3eb-43f8-aa12-2b7feea78f6a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.223301 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/084860df-b3eb-43f8-aa12-2b7feea78f6a-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dqxfw\" (UID: \"084860df-b3eb-43f8-aa12-2b7feea78f6a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.223365 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/084860df-b3eb-43f8-aa12-2b7feea78f6a-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dqxfw\" (UID: \"084860df-b3eb-43f8-aa12-2b7feea78f6a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.223426 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/084860df-b3eb-43f8-aa12-2b7feea78f6a-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dqxfw\" (UID: \"084860df-b3eb-43f8-aa12-2b7feea78f6a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.223484 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87gvp\" (UniqueName: \"kubernetes.io/projected/084860df-b3eb-43f8-aa12-2b7feea78f6a-kube-api-access-87gvp\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dqxfw\" (UID: \"084860df-b3eb-43f8-aa12-2b7feea78f6a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.223541 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/084860df-b3eb-43f8-aa12-2b7feea78f6a-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dqxfw\" (UID: \"084860df-b3eb-43f8-aa12-2b7feea78f6a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.224774 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/084860df-b3eb-43f8-aa12-2b7feea78f6a-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dqxfw\" (UID: \"084860df-b3eb-43f8-aa12-2b7feea78f6a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.229452 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/084860df-b3eb-43f8-aa12-2b7feea78f6a-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dqxfw\" (UID: \"084860df-b3eb-43f8-aa12-2b7feea78f6a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.229771 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/084860df-b3eb-43f8-aa12-2b7feea78f6a-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dqxfw\" (UID: \"084860df-b3eb-43f8-aa12-2b7feea78f6a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.234781 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/084860df-b3eb-43f8-aa12-2b7feea78f6a-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dqxfw\" (UID: \"084860df-b3eb-43f8-aa12-2b7feea78f6a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.245354 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87gvp\" (UniqueName: \"kubernetes.io/projected/084860df-b3eb-43f8-aa12-2b7feea78f6a-kube-api-access-87gvp\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-dqxfw\" (UID: \"084860df-b3eb-43f8-aa12-2b7feea78f6a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.362303 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.888438 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw"] Oct 03 13:31:45 crc kubenswrapper[4578]: I1003 13:31:45.888940 4578 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 13:31:46 crc kubenswrapper[4578]: I1003 13:31:46.737232 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" event={"ID":"084860df-b3eb-43f8-aa12-2b7feea78f6a","Type":"ContainerStarted","Data":"6d0db6c5c4c22d76a3033379c33b59804905550aa27f29ad2861c2da1107047a"} Oct 03 13:31:46 crc kubenswrapper[4578]: I1003 13:31:46.737557 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" event={"ID":"084860df-b3eb-43f8-aa12-2b7feea78f6a","Type":"ContainerStarted","Data":"49ec6cbd45d775022cc455f00606a453636e2503c653ba7a0ad6025a451dcedf"} Oct 03 13:31:46 crc kubenswrapper[4578]: I1003 13:31:46.760055 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" podStartSLOduration=1.6136902850000001 podStartE2EDuration="1.760038337s" podCreationTimestamp="2025-10-03 13:31:45 +0000 UTC" firstStartedPulling="2025-10-03 13:31:45.888610533 +0000 UTC m=+2441.687082727" lastFinishedPulling="2025-10-03 13:31:46.034958595 +0000 UTC m=+2441.833430779" observedRunningTime="2025-10-03 13:31:46.753408999 +0000 UTC m=+2442.551881183" watchObservedRunningTime="2025-10-03 13:31:46.760038337 +0000 UTC m=+2442.558510521" Oct 03 13:31:51 crc kubenswrapper[4578]: I1003 13:31:51.908862 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:31:51 crc kubenswrapper[4578]: E1003 13:31:51.909708 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:32:02 crc kubenswrapper[4578]: I1003 13:32:02.909190 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:32:02 crc kubenswrapper[4578]: E1003 13:32:02.910106 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:32:16 crc kubenswrapper[4578]: I1003 13:32:16.909160 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:32:16 crc kubenswrapper[4578]: E1003 13:32:16.909942 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:32:19 crc kubenswrapper[4578]: I1003 13:32:19.009385 4578 generic.go:334] "Generic (PLEG): container finished" podID="084860df-b3eb-43f8-aa12-2b7feea78f6a" containerID="6d0db6c5c4c22d76a3033379c33b59804905550aa27f29ad2861c2da1107047a" exitCode=2 Oct 03 13:32:19 crc kubenswrapper[4578]: I1003 13:32:19.009424 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" event={"ID":"084860df-b3eb-43f8-aa12-2b7feea78f6a","Type":"ContainerDied","Data":"6d0db6c5c4c22d76a3033379c33b59804905550aa27f29ad2861c2da1107047a"} Oct 03 13:32:20 crc kubenswrapper[4578]: I1003 13:32:20.396438 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" Oct 03 13:32:20 crc kubenswrapper[4578]: I1003 13:32:20.523074 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/084860df-b3eb-43f8-aa12-2b7feea78f6a-ssh-key\") pod \"084860df-b3eb-43f8-aa12-2b7feea78f6a\" (UID: \"084860df-b3eb-43f8-aa12-2b7feea78f6a\") " Oct 03 13:32:20 crc kubenswrapper[4578]: I1003 13:32:20.523190 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/084860df-b3eb-43f8-aa12-2b7feea78f6a-ovn-combined-ca-bundle\") pod \"084860df-b3eb-43f8-aa12-2b7feea78f6a\" (UID: \"084860df-b3eb-43f8-aa12-2b7feea78f6a\") " Oct 03 13:32:20 crc kubenswrapper[4578]: I1003 13:32:20.523268 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/084860df-b3eb-43f8-aa12-2b7feea78f6a-ovncontroller-config-0\") pod \"084860df-b3eb-43f8-aa12-2b7feea78f6a\" (UID: \"084860df-b3eb-43f8-aa12-2b7feea78f6a\") " Oct 03 13:32:20 crc kubenswrapper[4578]: I1003 13:32:20.523310 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/084860df-b3eb-43f8-aa12-2b7feea78f6a-inventory\") pod \"084860df-b3eb-43f8-aa12-2b7feea78f6a\" (UID: \"084860df-b3eb-43f8-aa12-2b7feea78f6a\") " Oct 03 13:32:20 crc kubenswrapper[4578]: I1003 13:32:20.523444 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87gvp\" (UniqueName: \"kubernetes.io/projected/084860df-b3eb-43f8-aa12-2b7feea78f6a-kube-api-access-87gvp\") pod \"084860df-b3eb-43f8-aa12-2b7feea78f6a\" (UID: \"084860df-b3eb-43f8-aa12-2b7feea78f6a\") " Oct 03 13:32:20 crc kubenswrapper[4578]: I1003 13:32:20.529229 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/084860df-b3eb-43f8-aa12-2b7feea78f6a-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "084860df-b3eb-43f8-aa12-2b7feea78f6a" (UID: "084860df-b3eb-43f8-aa12-2b7feea78f6a"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:32:20 crc kubenswrapper[4578]: I1003 13:32:20.568661 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/084860df-b3eb-43f8-aa12-2b7feea78f6a-kube-api-access-87gvp" (OuterVolumeSpecName: "kube-api-access-87gvp") pod "084860df-b3eb-43f8-aa12-2b7feea78f6a" (UID: "084860df-b3eb-43f8-aa12-2b7feea78f6a"). InnerVolumeSpecName "kube-api-access-87gvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:32:20 crc kubenswrapper[4578]: I1003 13:32:20.569273 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/084860df-b3eb-43f8-aa12-2b7feea78f6a-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "084860df-b3eb-43f8-aa12-2b7feea78f6a" (UID: "084860df-b3eb-43f8-aa12-2b7feea78f6a"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:32:20 crc kubenswrapper[4578]: I1003 13:32:20.574145 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/084860df-b3eb-43f8-aa12-2b7feea78f6a-inventory" (OuterVolumeSpecName: "inventory") pod "084860df-b3eb-43f8-aa12-2b7feea78f6a" (UID: "084860df-b3eb-43f8-aa12-2b7feea78f6a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:32:20 crc kubenswrapper[4578]: I1003 13:32:20.576443 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/084860df-b3eb-43f8-aa12-2b7feea78f6a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "084860df-b3eb-43f8-aa12-2b7feea78f6a" (UID: "084860df-b3eb-43f8-aa12-2b7feea78f6a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:32:20 crc kubenswrapper[4578]: I1003 13:32:20.625454 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87gvp\" (UniqueName: \"kubernetes.io/projected/084860df-b3eb-43f8-aa12-2b7feea78f6a-kube-api-access-87gvp\") on node \"crc\" DevicePath \"\"" Oct 03 13:32:20 crc kubenswrapper[4578]: I1003 13:32:20.625499 4578 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/084860df-b3eb-43f8-aa12-2b7feea78f6a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:32:20 crc kubenswrapper[4578]: I1003 13:32:20.625510 4578 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/084860df-b3eb-43f8-aa12-2b7feea78f6a-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:32:20 crc kubenswrapper[4578]: I1003 13:32:20.625520 4578 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/084860df-b3eb-43f8-aa12-2b7feea78f6a-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:32:20 crc kubenswrapper[4578]: I1003 13:32:20.625530 4578 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/084860df-b3eb-43f8-aa12-2b7feea78f6a-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 13:32:21 crc kubenswrapper[4578]: I1003 13:32:21.040237 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" event={"ID":"084860df-b3eb-43f8-aa12-2b7feea78f6a","Type":"ContainerDied","Data":"49ec6cbd45d775022cc455f00606a453636e2503c653ba7a0ad6025a451dcedf"} Oct 03 13:32:21 crc kubenswrapper[4578]: I1003 13:32:21.040284 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49ec6cbd45d775022cc455f00606a453636e2503c653ba7a0ad6025a451dcedf" Oct 03 13:32:21 crc kubenswrapper[4578]: I1003 13:32:21.040290 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-dqxfw" Oct 03 13:32:27 crc kubenswrapper[4578]: I1003 13:32:27.909382 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:32:27 crc kubenswrapper[4578]: E1003 13:32:27.910175 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:32:38 crc kubenswrapper[4578]: I1003 13:32:38.908960 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:32:39 crc kubenswrapper[4578]: I1003 13:32:39.197072 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"32a0b82351869c87347f7c4f14ff5c12356fd24347e53a70e720916381ba5d60"} Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.033873 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh"] Oct 03 13:33:38 crc kubenswrapper[4578]: E1003 13:33:38.034824 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="084860df-b3eb-43f8-aa12-2b7feea78f6a" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.034838 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="084860df-b3eb-43f8-aa12-2b7feea78f6a" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.035114 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="084860df-b3eb-43f8-aa12-2b7feea78f6a" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.035747 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.038469 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.038743 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.038853 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-442mn" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.038866 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.040149 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.050591 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh"] Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.172661 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f394e983-821e-4983-a093-711f2d6e0a23-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wn5jh\" (UID: \"f394e983-821e-4983-a093-711f2d6e0a23\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.172743 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f394e983-821e-4983-a093-711f2d6e0a23-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wn5jh\" (UID: \"f394e983-821e-4983-a093-711f2d6e0a23\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.172804 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f394e983-821e-4983-a093-711f2d6e0a23-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wn5jh\" (UID: \"f394e983-821e-4983-a093-711f2d6e0a23\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.172931 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f394e983-821e-4983-a093-711f2d6e0a23-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wn5jh\" (UID: \"f394e983-821e-4983-a093-711f2d6e0a23\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.172958 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgr6z\" (UniqueName: \"kubernetes.io/projected/f394e983-821e-4983-a093-711f2d6e0a23-kube-api-access-lgr6z\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wn5jh\" (UID: \"f394e983-821e-4983-a093-711f2d6e0a23\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.274693 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f394e983-821e-4983-a093-711f2d6e0a23-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wn5jh\" (UID: \"f394e983-821e-4983-a093-711f2d6e0a23\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.274752 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f394e983-821e-4983-a093-711f2d6e0a23-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wn5jh\" (UID: \"f394e983-821e-4983-a093-711f2d6e0a23\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.274777 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f394e983-821e-4983-a093-711f2d6e0a23-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wn5jh\" (UID: \"f394e983-821e-4983-a093-711f2d6e0a23\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.274837 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f394e983-821e-4983-a093-711f2d6e0a23-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wn5jh\" (UID: \"f394e983-821e-4983-a093-711f2d6e0a23\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.274856 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgr6z\" (UniqueName: \"kubernetes.io/projected/f394e983-821e-4983-a093-711f2d6e0a23-kube-api-access-lgr6z\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wn5jh\" (UID: \"f394e983-821e-4983-a093-711f2d6e0a23\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.279601 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f394e983-821e-4983-a093-711f2d6e0a23-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wn5jh\" (UID: \"f394e983-821e-4983-a093-711f2d6e0a23\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.299488 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f394e983-821e-4983-a093-711f2d6e0a23-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wn5jh\" (UID: \"f394e983-821e-4983-a093-711f2d6e0a23\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.313679 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f394e983-821e-4983-a093-711f2d6e0a23-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wn5jh\" (UID: \"f394e983-821e-4983-a093-711f2d6e0a23\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.316101 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f394e983-821e-4983-a093-711f2d6e0a23-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wn5jh\" (UID: \"f394e983-821e-4983-a093-711f2d6e0a23\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.318440 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgr6z\" (UniqueName: \"kubernetes.io/projected/f394e983-821e-4983-a093-711f2d6e0a23-kube-api-access-lgr6z\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wn5jh\" (UID: \"f394e983-821e-4983-a093-711f2d6e0a23\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.356436 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" Oct 03 13:33:38 crc kubenswrapper[4578]: I1003 13:33:38.920288 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh"] Oct 03 13:33:39 crc kubenswrapper[4578]: I1003 13:33:39.723882 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" event={"ID":"f394e983-821e-4983-a093-711f2d6e0a23","Type":"ContainerStarted","Data":"10d4437ae760cf581feda9b73f1cfdf6a3e6e9e42e5bdc6ac15e4ce4bec0de05"} Oct 03 13:33:39 crc kubenswrapper[4578]: I1003 13:33:39.724238 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" event={"ID":"f394e983-821e-4983-a093-711f2d6e0a23","Type":"ContainerStarted","Data":"e116b65d4ef1981298e1b093dd9d1238d58529a4e38125e53297789fddcbc0d9"} Oct 03 13:33:39 crc kubenswrapper[4578]: I1003 13:33:39.741977 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" podStartSLOduration=1.4492646759999999 podStartE2EDuration="1.741958292s" podCreationTimestamp="2025-10-03 13:33:38 +0000 UTC" firstStartedPulling="2025-10-03 13:33:38.918867175 +0000 UTC m=+2554.717339359" lastFinishedPulling="2025-10-03 13:33:39.211560791 +0000 UTC m=+2555.010032975" observedRunningTime="2025-10-03 13:33:39.737432041 +0000 UTC m=+2555.535904225" watchObservedRunningTime="2025-10-03 13:33:39.741958292 +0000 UTC m=+2555.540430476" Oct 03 13:34:12 crc kubenswrapper[4578]: E1003 13:34:12.464663 4578 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf394e983_821e_4983_a093_711f2d6e0a23.slice/crio-10d4437ae760cf581feda9b73f1cfdf6a3e6e9e42e5bdc6ac15e4ce4bec0de05.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf394e983_821e_4983_a093_711f2d6e0a23.slice/crio-conmon-10d4437ae760cf581feda9b73f1cfdf6a3e6e9e42e5bdc6ac15e4ce4bec0de05.scope\": RecentStats: unable to find data in memory cache]" Oct 03 13:34:13 crc kubenswrapper[4578]: I1003 13:34:13.021809 4578 generic.go:334] "Generic (PLEG): container finished" podID="f394e983-821e-4983-a093-711f2d6e0a23" containerID="10d4437ae760cf581feda9b73f1cfdf6a3e6e9e42e5bdc6ac15e4ce4bec0de05" exitCode=2 Oct 03 13:34:13 crc kubenswrapper[4578]: I1003 13:34:13.021914 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" event={"ID":"f394e983-821e-4983-a093-711f2d6e0a23","Type":"ContainerDied","Data":"10d4437ae760cf581feda9b73f1cfdf6a3e6e9e42e5bdc6ac15e4ce4bec0de05"} Oct 03 13:34:14 crc kubenswrapper[4578]: I1003 13:34:14.437766 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" Oct 03 13:34:14 crc kubenswrapper[4578]: I1003 13:34:14.626297 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f394e983-821e-4983-a093-711f2d6e0a23-inventory\") pod \"f394e983-821e-4983-a093-711f2d6e0a23\" (UID: \"f394e983-821e-4983-a093-711f2d6e0a23\") " Oct 03 13:34:14 crc kubenswrapper[4578]: I1003 13:34:14.627389 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f394e983-821e-4983-a093-711f2d6e0a23-ovn-combined-ca-bundle\") pod \"f394e983-821e-4983-a093-711f2d6e0a23\" (UID: \"f394e983-821e-4983-a093-711f2d6e0a23\") " Oct 03 13:34:14 crc kubenswrapper[4578]: I1003 13:34:14.627444 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgr6z\" (UniqueName: \"kubernetes.io/projected/f394e983-821e-4983-a093-711f2d6e0a23-kube-api-access-lgr6z\") pod \"f394e983-821e-4983-a093-711f2d6e0a23\" (UID: \"f394e983-821e-4983-a093-711f2d6e0a23\") " Oct 03 13:34:14 crc kubenswrapper[4578]: I1003 13:34:14.627467 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f394e983-821e-4983-a093-711f2d6e0a23-ovncontroller-config-0\") pod \"f394e983-821e-4983-a093-711f2d6e0a23\" (UID: \"f394e983-821e-4983-a093-711f2d6e0a23\") " Oct 03 13:34:14 crc kubenswrapper[4578]: I1003 13:34:14.627489 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f394e983-821e-4983-a093-711f2d6e0a23-ssh-key\") pod \"f394e983-821e-4983-a093-711f2d6e0a23\" (UID: \"f394e983-821e-4983-a093-711f2d6e0a23\") " Oct 03 13:34:14 crc kubenswrapper[4578]: I1003 13:34:14.634619 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f394e983-821e-4983-a093-711f2d6e0a23-kube-api-access-lgr6z" (OuterVolumeSpecName: "kube-api-access-lgr6z") pod "f394e983-821e-4983-a093-711f2d6e0a23" (UID: "f394e983-821e-4983-a093-711f2d6e0a23"). InnerVolumeSpecName "kube-api-access-lgr6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:34:14 crc kubenswrapper[4578]: I1003 13:34:14.635380 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f394e983-821e-4983-a093-711f2d6e0a23-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "f394e983-821e-4983-a093-711f2d6e0a23" (UID: "f394e983-821e-4983-a093-711f2d6e0a23"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:34:14 crc kubenswrapper[4578]: I1003 13:34:14.654439 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f394e983-821e-4983-a093-711f2d6e0a23-inventory" (OuterVolumeSpecName: "inventory") pod "f394e983-821e-4983-a093-711f2d6e0a23" (UID: "f394e983-821e-4983-a093-711f2d6e0a23"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:34:14 crc kubenswrapper[4578]: I1003 13:34:14.654516 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f394e983-821e-4983-a093-711f2d6e0a23-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "f394e983-821e-4983-a093-711f2d6e0a23" (UID: "f394e983-821e-4983-a093-711f2d6e0a23"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:34:14 crc kubenswrapper[4578]: I1003 13:34:14.658381 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f394e983-821e-4983-a093-711f2d6e0a23-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f394e983-821e-4983-a093-711f2d6e0a23" (UID: "f394e983-821e-4983-a093-711f2d6e0a23"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:34:14 crc kubenswrapper[4578]: I1003 13:34:14.729883 4578 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f394e983-821e-4983-a093-711f2d6e0a23-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:14 crc kubenswrapper[4578]: I1003 13:34:14.729941 4578 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f394e983-821e-4983-a093-711f2d6e0a23-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:14 crc kubenswrapper[4578]: I1003 13:34:14.729954 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgr6z\" (UniqueName: \"kubernetes.io/projected/f394e983-821e-4983-a093-711f2d6e0a23-kube-api-access-lgr6z\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:14 crc kubenswrapper[4578]: I1003 13:34:14.729963 4578 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f394e983-821e-4983-a093-711f2d6e0a23-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:14 crc kubenswrapper[4578]: I1003 13:34:14.729971 4578 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f394e983-821e-4983-a093-711f2d6e0a23-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:34:15 crc kubenswrapper[4578]: I1003 13:34:15.040152 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" event={"ID":"f394e983-821e-4983-a093-711f2d6e0a23","Type":"ContainerDied","Data":"e116b65d4ef1981298e1b093dd9d1238d58529a4e38125e53297789fddcbc0d9"} Oct 03 13:34:15 crc kubenswrapper[4578]: I1003 13:34:15.040187 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e116b65d4ef1981298e1b093dd9d1238d58529a4e38125e53297789fddcbc0d9" Oct 03 13:34:15 crc kubenswrapper[4578]: I1003 13:34:15.040221 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wn5jh" Oct 03 13:35:05 crc kubenswrapper[4578]: I1003 13:35:05.092229 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:35:05 crc kubenswrapper[4578]: I1003 13:35:05.092872 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:35:35 crc kubenswrapper[4578]: I1003 13:35:35.092136 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:35:35 crc kubenswrapper[4578]: I1003 13:35:35.092821 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:35:53 crc kubenswrapper[4578]: I1003 13:35:53.752760 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6n4ls"] Oct 03 13:35:53 crc kubenswrapper[4578]: E1003 13:35:53.754207 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f394e983-821e-4983-a093-711f2d6e0a23" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:35:53 crc kubenswrapper[4578]: I1003 13:35:53.754229 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="f394e983-821e-4983-a093-711f2d6e0a23" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:35:53 crc kubenswrapper[4578]: I1003 13:35:53.754565 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="f394e983-821e-4983-a093-711f2d6e0a23" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:35:53 crc kubenswrapper[4578]: I1003 13:35:53.755988 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6n4ls" Oct 03 13:35:53 crc kubenswrapper[4578]: I1003 13:35:53.789338 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6n4ls"] Oct 03 13:35:53 crc kubenswrapper[4578]: I1003 13:35:53.795904 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fecd676-6f90-482c-b5a9-5c0aa4f94553-catalog-content\") pod \"community-operators-6n4ls\" (UID: \"2fecd676-6f90-482c-b5a9-5c0aa4f94553\") " pod="openshift-marketplace/community-operators-6n4ls" Oct 03 13:35:53 crc kubenswrapper[4578]: I1003 13:35:53.796078 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvvzd\" (UniqueName: \"kubernetes.io/projected/2fecd676-6f90-482c-b5a9-5c0aa4f94553-kube-api-access-pvvzd\") pod \"community-operators-6n4ls\" (UID: \"2fecd676-6f90-482c-b5a9-5c0aa4f94553\") " pod="openshift-marketplace/community-operators-6n4ls" Oct 03 13:35:53 crc kubenswrapper[4578]: I1003 13:35:53.796208 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fecd676-6f90-482c-b5a9-5c0aa4f94553-utilities\") pod \"community-operators-6n4ls\" (UID: \"2fecd676-6f90-482c-b5a9-5c0aa4f94553\") " pod="openshift-marketplace/community-operators-6n4ls" Oct 03 13:35:53 crc kubenswrapper[4578]: I1003 13:35:53.897910 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvvzd\" (UniqueName: \"kubernetes.io/projected/2fecd676-6f90-482c-b5a9-5c0aa4f94553-kube-api-access-pvvzd\") pod \"community-operators-6n4ls\" (UID: \"2fecd676-6f90-482c-b5a9-5c0aa4f94553\") " pod="openshift-marketplace/community-operators-6n4ls" Oct 03 13:35:53 crc kubenswrapper[4578]: I1003 13:35:53.898325 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fecd676-6f90-482c-b5a9-5c0aa4f94553-utilities\") pod \"community-operators-6n4ls\" (UID: \"2fecd676-6f90-482c-b5a9-5c0aa4f94553\") " pod="openshift-marketplace/community-operators-6n4ls" Oct 03 13:35:53 crc kubenswrapper[4578]: I1003 13:35:53.898566 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fecd676-6f90-482c-b5a9-5c0aa4f94553-catalog-content\") pod \"community-operators-6n4ls\" (UID: \"2fecd676-6f90-482c-b5a9-5c0aa4f94553\") " pod="openshift-marketplace/community-operators-6n4ls" Oct 03 13:35:53 crc kubenswrapper[4578]: I1003 13:35:53.899059 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fecd676-6f90-482c-b5a9-5c0aa4f94553-catalog-content\") pod \"community-operators-6n4ls\" (UID: \"2fecd676-6f90-482c-b5a9-5c0aa4f94553\") " pod="openshift-marketplace/community-operators-6n4ls" Oct 03 13:35:53 crc kubenswrapper[4578]: I1003 13:35:53.899333 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fecd676-6f90-482c-b5a9-5c0aa4f94553-utilities\") pod \"community-operators-6n4ls\" (UID: \"2fecd676-6f90-482c-b5a9-5c0aa4f94553\") " pod="openshift-marketplace/community-operators-6n4ls" Oct 03 13:35:53 crc kubenswrapper[4578]: I1003 13:35:53.917010 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvvzd\" (UniqueName: \"kubernetes.io/projected/2fecd676-6f90-482c-b5a9-5c0aa4f94553-kube-api-access-pvvzd\") pod \"community-operators-6n4ls\" (UID: \"2fecd676-6f90-482c-b5a9-5c0aa4f94553\") " pod="openshift-marketplace/community-operators-6n4ls" Oct 03 13:35:54 crc kubenswrapper[4578]: I1003 13:35:54.078684 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6n4ls" Oct 03 13:35:54 crc kubenswrapper[4578]: I1003 13:35:54.701805 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6n4ls"] Oct 03 13:35:54 crc kubenswrapper[4578]: I1003 13:35:54.896264 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6n4ls" event={"ID":"2fecd676-6f90-482c-b5a9-5c0aa4f94553","Type":"ContainerStarted","Data":"ed9fd443b0db9dd3965c78ef4fe06dc2aee2f17c6bffa471a711a3320cb6e555"} Oct 03 13:35:55 crc kubenswrapper[4578]: I1003 13:35:55.906427 4578 generic.go:334] "Generic (PLEG): container finished" podID="2fecd676-6f90-482c-b5a9-5c0aa4f94553" containerID="0bc0de0c98f46381d0df99e2389dbb845198c812fe694e3aa92e41fb331e9626" exitCode=0 Oct 03 13:35:55 crc kubenswrapper[4578]: I1003 13:35:55.906476 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6n4ls" event={"ID":"2fecd676-6f90-482c-b5a9-5c0aa4f94553","Type":"ContainerDied","Data":"0bc0de0c98f46381d0df99e2389dbb845198c812fe694e3aa92e41fb331e9626"} Oct 03 13:35:56 crc kubenswrapper[4578]: I1003 13:35:56.925318 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6n4ls" event={"ID":"2fecd676-6f90-482c-b5a9-5c0aa4f94553","Type":"ContainerStarted","Data":"4177755c23bb94ced0908ce97fdab60032f5f6fb59f4862d407647e9702d9032"} Oct 03 13:35:58 crc kubenswrapper[4578]: I1003 13:35:58.954705 4578 generic.go:334] "Generic (PLEG): container finished" podID="2fecd676-6f90-482c-b5a9-5c0aa4f94553" containerID="4177755c23bb94ced0908ce97fdab60032f5f6fb59f4862d407647e9702d9032" exitCode=0 Oct 03 13:35:58 crc kubenswrapper[4578]: I1003 13:35:58.954799 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6n4ls" event={"ID":"2fecd676-6f90-482c-b5a9-5c0aa4f94553","Type":"ContainerDied","Data":"4177755c23bb94ced0908ce97fdab60032f5f6fb59f4862d407647e9702d9032"} Oct 03 13:35:59 crc kubenswrapper[4578]: I1003 13:35:59.990335 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6n4ls" event={"ID":"2fecd676-6f90-482c-b5a9-5c0aa4f94553","Type":"ContainerStarted","Data":"2ebfc39cabad8df71cd30128c9e27b3c8acc98d16e9616835e76170910f03658"} Oct 03 13:36:00 crc kubenswrapper[4578]: I1003 13:36:00.051006 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6n4ls" podStartSLOduration=3.271376838 podStartE2EDuration="7.050983022s" podCreationTimestamp="2025-10-03 13:35:53 +0000 UTC" firstStartedPulling="2025-10-03 13:35:55.908876677 +0000 UTC m=+2691.707348861" lastFinishedPulling="2025-10-03 13:35:59.688482861 +0000 UTC m=+2695.486955045" observedRunningTime="2025-10-03 13:36:00.022595008 +0000 UTC m=+2695.821067192" watchObservedRunningTime="2025-10-03 13:36:00.050983022 +0000 UTC m=+2695.849455206" Oct 03 13:36:04 crc kubenswrapper[4578]: I1003 13:36:04.079511 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6n4ls" Oct 03 13:36:04 crc kubenswrapper[4578]: I1003 13:36:04.080069 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6n4ls" Oct 03 13:36:04 crc kubenswrapper[4578]: I1003 13:36:04.127542 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6n4ls" Oct 03 13:36:05 crc kubenswrapper[4578]: I1003 13:36:05.078567 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6n4ls" Oct 03 13:36:05 crc kubenswrapper[4578]: I1003 13:36:05.091084 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:36:05 crc kubenswrapper[4578]: I1003 13:36:05.091174 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:36:05 crc kubenswrapper[4578]: I1003 13:36:05.091692 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 13:36:05 crc kubenswrapper[4578]: I1003 13:36:05.092597 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"32a0b82351869c87347f7c4f14ff5c12356fd24347e53a70e720916381ba5d60"} pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:36:05 crc kubenswrapper[4578]: I1003 13:36:05.092705 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" containerID="cri-o://32a0b82351869c87347f7c4f14ff5c12356fd24347e53a70e720916381ba5d60" gracePeriod=600 Oct 03 13:36:05 crc kubenswrapper[4578]: I1003 13:36:05.139229 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6n4ls"] Oct 03 13:36:05 crc kubenswrapper[4578]: E1003 13:36:05.171103 4578 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7d7e663_26ba_48e2_849e_f7c9554fd7a7.slice/crio-32a0b82351869c87347f7c4f14ff5c12356fd24347e53a70e720916381ba5d60.scope\": RecentStats: unable to find data in memory cache]" Oct 03 13:36:06 crc kubenswrapper[4578]: I1003 13:36:06.042992 4578 generic.go:334] "Generic (PLEG): container finished" podID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerID="32a0b82351869c87347f7c4f14ff5c12356fd24347e53a70e720916381ba5d60" exitCode=0 Oct 03 13:36:06 crc kubenswrapper[4578]: I1003 13:36:06.043069 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerDied","Data":"32a0b82351869c87347f7c4f14ff5c12356fd24347e53a70e720916381ba5d60"} Oct 03 13:36:06 crc kubenswrapper[4578]: I1003 13:36:06.043239 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406"} Oct 03 13:36:06 crc kubenswrapper[4578]: I1003 13:36:06.043255 4578 scope.go:117] "RemoveContainer" containerID="29c7f118f1f5be4ca10c120d31f514cf4d324f5b4280e0ea2c8eaaae3c0e48e7" Oct 03 13:36:07 crc kubenswrapper[4578]: I1003 13:36:07.058127 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6n4ls" podUID="2fecd676-6f90-482c-b5a9-5c0aa4f94553" containerName="registry-server" containerID="cri-o://2ebfc39cabad8df71cd30128c9e27b3c8acc98d16e9616835e76170910f03658" gracePeriod=2 Oct 03 13:36:07 crc kubenswrapper[4578]: I1003 13:36:07.489026 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6n4ls" Oct 03 13:36:07 crc kubenswrapper[4578]: I1003 13:36:07.598981 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvvzd\" (UniqueName: \"kubernetes.io/projected/2fecd676-6f90-482c-b5a9-5c0aa4f94553-kube-api-access-pvvzd\") pod \"2fecd676-6f90-482c-b5a9-5c0aa4f94553\" (UID: \"2fecd676-6f90-482c-b5a9-5c0aa4f94553\") " Oct 03 13:36:07 crc kubenswrapper[4578]: I1003 13:36:07.599049 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fecd676-6f90-482c-b5a9-5c0aa4f94553-utilities\") pod \"2fecd676-6f90-482c-b5a9-5c0aa4f94553\" (UID: \"2fecd676-6f90-482c-b5a9-5c0aa4f94553\") " Oct 03 13:36:07 crc kubenswrapper[4578]: I1003 13:36:07.599197 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fecd676-6f90-482c-b5a9-5c0aa4f94553-catalog-content\") pod \"2fecd676-6f90-482c-b5a9-5c0aa4f94553\" (UID: \"2fecd676-6f90-482c-b5a9-5c0aa4f94553\") " Oct 03 13:36:07 crc kubenswrapper[4578]: I1003 13:36:07.599903 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fecd676-6f90-482c-b5a9-5c0aa4f94553-utilities" (OuterVolumeSpecName: "utilities") pod "2fecd676-6f90-482c-b5a9-5c0aa4f94553" (UID: "2fecd676-6f90-482c-b5a9-5c0aa4f94553"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:36:07 crc kubenswrapper[4578]: I1003 13:36:07.605743 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fecd676-6f90-482c-b5a9-5c0aa4f94553-kube-api-access-pvvzd" (OuterVolumeSpecName: "kube-api-access-pvvzd") pod "2fecd676-6f90-482c-b5a9-5c0aa4f94553" (UID: "2fecd676-6f90-482c-b5a9-5c0aa4f94553"). InnerVolumeSpecName "kube-api-access-pvvzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:36:07 crc kubenswrapper[4578]: I1003 13:36:07.658041 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fecd676-6f90-482c-b5a9-5c0aa4f94553-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2fecd676-6f90-482c-b5a9-5c0aa4f94553" (UID: "2fecd676-6f90-482c-b5a9-5c0aa4f94553"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:36:07 crc kubenswrapper[4578]: I1003 13:36:07.701506 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fecd676-6f90-482c-b5a9-5c0aa4f94553-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:36:07 crc kubenswrapper[4578]: I1003 13:36:07.701547 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fecd676-6f90-482c-b5a9-5c0aa4f94553-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:36:07 crc kubenswrapper[4578]: I1003 13:36:07.701564 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvvzd\" (UniqueName: \"kubernetes.io/projected/2fecd676-6f90-482c-b5a9-5c0aa4f94553-kube-api-access-pvvzd\") on node \"crc\" DevicePath \"\"" Oct 03 13:36:08 crc kubenswrapper[4578]: I1003 13:36:08.071503 4578 generic.go:334] "Generic (PLEG): container finished" podID="2fecd676-6f90-482c-b5a9-5c0aa4f94553" containerID="2ebfc39cabad8df71cd30128c9e27b3c8acc98d16e9616835e76170910f03658" exitCode=0 Oct 03 13:36:08 crc kubenswrapper[4578]: I1003 13:36:08.071548 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6n4ls" Oct 03 13:36:08 crc kubenswrapper[4578]: I1003 13:36:08.071574 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6n4ls" event={"ID":"2fecd676-6f90-482c-b5a9-5c0aa4f94553","Type":"ContainerDied","Data":"2ebfc39cabad8df71cd30128c9e27b3c8acc98d16e9616835e76170910f03658"} Oct 03 13:36:08 crc kubenswrapper[4578]: I1003 13:36:08.071677 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6n4ls" event={"ID":"2fecd676-6f90-482c-b5a9-5c0aa4f94553","Type":"ContainerDied","Data":"ed9fd443b0db9dd3965c78ef4fe06dc2aee2f17c6bffa471a711a3320cb6e555"} Oct 03 13:36:08 crc kubenswrapper[4578]: I1003 13:36:08.071755 4578 scope.go:117] "RemoveContainer" containerID="2ebfc39cabad8df71cd30128c9e27b3c8acc98d16e9616835e76170910f03658" Oct 03 13:36:08 crc kubenswrapper[4578]: I1003 13:36:08.101041 4578 scope.go:117] "RemoveContainer" containerID="4177755c23bb94ced0908ce97fdab60032f5f6fb59f4862d407647e9702d9032" Oct 03 13:36:08 crc kubenswrapper[4578]: I1003 13:36:08.131260 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6n4ls"] Oct 03 13:36:08 crc kubenswrapper[4578]: I1003 13:36:08.137088 4578 scope.go:117] "RemoveContainer" containerID="0bc0de0c98f46381d0df99e2389dbb845198c812fe694e3aa92e41fb331e9626" Oct 03 13:36:08 crc kubenswrapper[4578]: I1003 13:36:08.140328 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6n4ls"] Oct 03 13:36:08 crc kubenswrapper[4578]: I1003 13:36:08.187383 4578 scope.go:117] "RemoveContainer" containerID="2ebfc39cabad8df71cd30128c9e27b3c8acc98d16e9616835e76170910f03658" Oct 03 13:36:08 crc kubenswrapper[4578]: E1003 13:36:08.187750 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ebfc39cabad8df71cd30128c9e27b3c8acc98d16e9616835e76170910f03658\": container with ID starting with 2ebfc39cabad8df71cd30128c9e27b3c8acc98d16e9616835e76170910f03658 not found: ID does not exist" containerID="2ebfc39cabad8df71cd30128c9e27b3c8acc98d16e9616835e76170910f03658" Oct 03 13:36:08 crc kubenswrapper[4578]: I1003 13:36:08.187782 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ebfc39cabad8df71cd30128c9e27b3c8acc98d16e9616835e76170910f03658"} err="failed to get container status \"2ebfc39cabad8df71cd30128c9e27b3c8acc98d16e9616835e76170910f03658\": rpc error: code = NotFound desc = could not find container \"2ebfc39cabad8df71cd30128c9e27b3c8acc98d16e9616835e76170910f03658\": container with ID starting with 2ebfc39cabad8df71cd30128c9e27b3c8acc98d16e9616835e76170910f03658 not found: ID does not exist" Oct 03 13:36:08 crc kubenswrapper[4578]: I1003 13:36:08.187808 4578 scope.go:117] "RemoveContainer" containerID="4177755c23bb94ced0908ce97fdab60032f5f6fb59f4862d407647e9702d9032" Oct 03 13:36:08 crc kubenswrapper[4578]: E1003 13:36:08.188103 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4177755c23bb94ced0908ce97fdab60032f5f6fb59f4862d407647e9702d9032\": container with ID starting with 4177755c23bb94ced0908ce97fdab60032f5f6fb59f4862d407647e9702d9032 not found: ID does not exist" containerID="4177755c23bb94ced0908ce97fdab60032f5f6fb59f4862d407647e9702d9032" Oct 03 13:36:08 crc kubenswrapper[4578]: I1003 13:36:08.188127 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4177755c23bb94ced0908ce97fdab60032f5f6fb59f4862d407647e9702d9032"} err="failed to get container status \"4177755c23bb94ced0908ce97fdab60032f5f6fb59f4862d407647e9702d9032\": rpc error: code = NotFound desc = could not find container \"4177755c23bb94ced0908ce97fdab60032f5f6fb59f4862d407647e9702d9032\": container with ID starting with 4177755c23bb94ced0908ce97fdab60032f5f6fb59f4862d407647e9702d9032 not found: ID does not exist" Oct 03 13:36:08 crc kubenswrapper[4578]: I1003 13:36:08.188145 4578 scope.go:117] "RemoveContainer" containerID="0bc0de0c98f46381d0df99e2389dbb845198c812fe694e3aa92e41fb331e9626" Oct 03 13:36:08 crc kubenswrapper[4578]: E1003 13:36:08.188355 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bc0de0c98f46381d0df99e2389dbb845198c812fe694e3aa92e41fb331e9626\": container with ID starting with 0bc0de0c98f46381d0df99e2389dbb845198c812fe694e3aa92e41fb331e9626 not found: ID does not exist" containerID="0bc0de0c98f46381d0df99e2389dbb845198c812fe694e3aa92e41fb331e9626" Oct 03 13:36:08 crc kubenswrapper[4578]: I1003 13:36:08.188382 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bc0de0c98f46381d0df99e2389dbb845198c812fe694e3aa92e41fb331e9626"} err="failed to get container status \"0bc0de0c98f46381d0df99e2389dbb845198c812fe694e3aa92e41fb331e9626\": rpc error: code = NotFound desc = could not find container \"0bc0de0c98f46381d0df99e2389dbb845198c812fe694e3aa92e41fb331e9626\": container with ID starting with 0bc0de0c98f46381d0df99e2389dbb845198c812fe694e3aa92e41fb331e9626 not found: ID does not exist" Oct 03 13:36:08 crc kubenswrapper[4578]: I1003 13:36:08.922610 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fecd676-6f90-482c-b5a9-5c0aa4f94553" path="/var/lib/kubelet/pods/2fecd676-6f90-482c-b5a9-5c0aa4f94553/volumes" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.033346 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6"] Oct 03 13:36:52 crc kubenswrapper[4578]: E1003 13:36:52.034306 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fecd676-6f90-482c-b5a9-5c0aa4f94553" containerName="extract-content" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.034323 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fecd676-6f90-482c-b5a9-5c0aa4f94553" containerName="extract-content" Oct 03 13:36:52 crc kubenswrapper[4578]: E1003 13:36:52.034339 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fecd676-6f90-482c-b5a9-5c0aa4f94553" containerName="extract-utilities" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.034347 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fecd676-6f90-482c-b5a9-5c0aa4f94553" containerName="extract-utilities" Oct 03 13:36:52 crc kubenswrapper[4578]: E1003 13:36:52.034358 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fecd676-6f90-482c-b5a9-5c0aa4f94553" containerName="registry-server" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.034366 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fecd676-6f90-482c-b5a9-5c0aa4f94553" containerName="registry-server" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.034620 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fecd676-6f90-482c-b5a9-5c0aa4f94553" containerName="registry-server" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.035355 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.037471 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.041761 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6"] Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.042801 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-442mn" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.042949 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.043582 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.043582 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.195343 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6ee1a77-57be-48c1-a61f-50719bfcbd29-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7ztl6\" (UID: \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.195406 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e6ee1a77-57be-48c1-a61f-50719bfcbd29-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7ztl6\" (UID: \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.195431 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6ee1a77-57be-48c1-a61f-50719bfcbd29-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7ztl6\" (UID: \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.195547 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c569t\" (UniqueName: \"kubernetes.io/projected/e6ee1a77-57be-48c1-a61f-50719bfcbd29-kube-api-access-c569t\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7ztl6\" (UID: \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.195732 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6ee1a77-57be-48c1-a61f-50719bfcbd29-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7ztl6\" (UID: \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.296763 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6ee1a77-57be-48c1-a61f-50719bfcbd29-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7ztl6\" (UID: \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.296801 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6ee1a77-57be-48c1-a61f-50719bfcbd29-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7ztl6\" (UID: \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.296839 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e6ee1a77-57be-48c1-a61f-50719bfcbd29-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7ztl6\" (UID: \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.296858 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6ee1a77-57be-48c1-a61f-50719bfcbd29-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7ztl6\" (UID: \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.296911 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c569t\" (UniqueName: \"kubernetes.io/projected/e6ee1a77-57be-48c1-a61f-50719bfcbd29-kube-api-access-c569t\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7ztl6\" (UID: \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.298171 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e6ee1a77-57be-48c1-a61f-50719bfcbd29-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7ztl6\" (UID: \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.307240 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6ee1a77-57be-48c1-a61f-50719bfcbd29-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7ztl6\" (UID: \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.307475 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6ee1a77-57be-48c1-a61f-50719bfcbd29-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7ztl6\" (UID: \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.309707 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6ee1a77-57be-48c1-a61f-50719bfcbd29-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7ztl6\" (UID: \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.314793 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c569t\" (UniqueName: \"kubernetes.io/projected/e6ee1a77-57be-48c1-a61f-50719bfcbd29-kube-api-access-c569t\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7ztl6\" (UID: \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.355364 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.929902 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6"] Oct 03 13:36:52 crc kubenswrapper[4578]: I1003 13:36:52.937147 4578 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 13:36:53 crc kubenswrapper[4578]: I1003 13:36:53.469194 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" event={"ID":"e6ee1a77-57be-48c1-a61f-50719bfcbd29","Type":"ContainerStarted","Data":"9b8dc242ae34d52b502c0672df796e35773948cec75a30e8c162b1cbe1674e3b"} Oct 03 13:36:54 crc kubenswrapper[4578]: I1003 13:36:54.481050 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" event={"ID":"e6ee1a77-57be-48c1-a61f-50719bfcbd29","Type":"ContainerStarted","Data":"aa66394aa10fbec4b87b93e4bd4f38c51b9f306e854210814f87bf571e79496d"} Oct 03 13:36:54 crc kubenswrapper[4578]: I1003 13:36:54.502905 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" podStartSLOduration=1.899301079 podStartE2EDuration="2.502884148s" podCreationTimestamp="2025-10-03 13:36:52 +0000 UTC" firstStartedPulling="2025-10-03 13:36:52.9368595 +0000 UTC m=+2748.735331695" lastFinishedPulling="2025-10-03 13:36:53.54044257 +0000 UTC m=+2749.338914764" observedRunningTime="2025-10-03 13:36:54.496541801 +0000 UTC m=+2750.295013985" watchObservedRunningTime="2025-10-03 13:36:54.502884148 +0000 UTC m=+2750.301356332" Oct 03 13:37:22 crc kubenswrapper[4578]: I1003 13:37:22.122341 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-btppq"] Oct 03 13:37:22 crc kubenswrapper[4578]: I1003 13:37:22.125437 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-btppq" Oct 03 13:37:22 crc kubenswrapper[4578]: I1003 13:37:22.132663 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-btppq"] Oct 03 13:37:22 crc kubenswrapper[4578]: I1003 13:37:22.313455 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a3d676d-e014-4e02-b777-3a49b0ff11d1-catalog-content\") pod \"certified-operators-btppq\" (UID: \"3a3d676d-e014-4e02-b777-3a49b0ff11d1\") " pod="openshift-marketplace/certified-operators-btppq" Oct 03 13:37:22 crc kubenswrapper[4578]: I1003 13:37:22.313516 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmn8p\" (UniqueName: \"kubernetes.io/projected/3a3d676d-e014-4e02-b777-3a49b0ff11d1-kube-api-access-mmn8p\") pod \"certified-operators-btppq\" (UID: \"3a3d676d-e014-4e02-b777-3a49b0ff11d1\") " pod="openshift-marketplace/certified-operators-btppq" Oct 03 13:37:22 crc kubenswrapper[4578]: I1003 13:37:22.313749 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a3d676d-e014-4e02-b777-3a49b0ff11d1-utilities\") pod \"certified-operators-btppq\" (UID: \"3a3d676d-e014-4e02-b777-3a49b0ff11d1\") " pod="openshift-marketplace/certified-operators-btppq" Oct 03 13:37:22 crc kubenswrapper[4578]: I1003 13:37:22.415870 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a3d676d-e014-4e02-b777-3a49b0ff11d1-catalog-content\") pod \"certified-operators-btppq\" (UID: \"3a3d676d-e014-4e02-b777-3a49b0ff11d1\") " pod="openshift-marketplace/certified-operators-btppq" Oct 03 13:37:22 crc kubenswrapper[4578]: I1003 13:37:22.415914 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmn8p\" (UniqueName: \"kubernetes.io/projected/3a3d676d-e014-4e02-b777-3a49b0ff11d1-kube-api-access-mmn8p\") pod \"certified-operators-btppq\" (UID: \"3a3d676d-e014-4e02-b777-3a49b0ff11d1\") " pod="openshift-marketplace/certified-operators-btppq" Oct 03 13:37:22 crc kubenswrapper[4578]: I1003 13:37:22.416010 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a3d676d-e014-4e02-b777-3a49b0ff11d1-utilities\") pod \"certified-operators-btppq\" (UID: \"3a3d676d-e014-4e02-b777-3a49b0ff11d1\") " pod="openshift-marketplace/certified-operators-btppq" Oct 03 13:37:22 crc kubenswrapper[4578]: I1003 13:37:22.416491 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a3d676d-e014-4e02-b777-3a49b0ff11d1-catalog-content\") pod \"certified-operators-btppq\" (UID: \"3a3d676d-e014-4e02-b777-3a49b0ff11d1\") " pod="openshift-marketplace/certified-operators-btppq" Oct 03 13:37:22 crc kubenswrapper[4578]: I1003 13:37:22.416506 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a3d676d-e014-4e02-b777-3a49b0ff11d1-utilities\") pod \"certified-operators-btppq\" (UID: \"3a3d676d-e014-4e02-b777-3a49b0ff11d1\") " pod="openshift-marketplace/certified-operators-btppq" Oct 03 13:37:22 crc kubenswrapper[4578]: I1003 13:37:22.437307 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmn8p\" (UniqueName: \"kubernetes.io/projected/3a3d676d-e014-4e02-b777-3a49b0ff11d1-kube-api-access-mmn8p\") pod \"certified-operators-btppq\" (UID: \"3a3d676d-e014-4e02-b777-3a49b0ff11d1\") " pod="openshift-marketplace/certified-operators-btppq" Oct 03 13:37:22 crc kubenswrapper[4578]: I1003 13:37:22.451008 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-btppq" Oct 03 13:37:23 crc kubenswrapper[4578]: I1003 13:37:23.095268 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-btppq"] Oct 03 13:37:23 crc kubenswrapper[4578]: W1003 13:37:23.105118 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a3d676d_e014_4e02_b777_3a49b0ff11d1.slice/crio-a69eb8075b58731f2e19a581fc157cd1ee2366a36ddb2b98b1ea81661594adb9 WatchSource:0}: Error finding container a69eb8075b58731f2e19a581fc157cd1ee2366a36ddb2b98b1ea81661594adb9: Status 404 returned error can't find the container with id a69eb8075b58731f2e19a581fc157cd1ee2366a36ddb2b98b1ea81661594adb9 Oct 03 13:37:23 crc kubenswrapper[4578]: I1003 13:37:23.740533 4578 generic.go:334] "Generic (PLEG): container finished" podID="3a3d676d-e014-4e02-b777-3a49b0ff11d1" containerID="8ed1f421bbeb8edb6464024799a5b45584846ce8690c812ad2c4404ac690b8ee" exitCode=0 Oct 03 13:37:23 crc kubenswrapper[4578]: I1003 13:37:23.740696 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-btppq" event={"ID":"3a3d676d-e014-4e02-b777-3a49b0ff11d1","Type":"ContainerDied","Data":"8ed1f421bbeb8edb6464024799a5b45584846ce8690c812ad2c4404ac690b8ee"} Oct 03 13:37:23 crc kubenswrapper[4578]: I1003 13:37:23.740847 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-btppq" event={"ID":"3a3d676d-e014-4e02-b777-3a49b0ff11d1","Type":"ContainerStarted","Data":"a69eb8075b58731f2e19a581fc157cd1ee2366a36ddb2b98b1ea81661594adb9"} Oct 03 13:37:24 crc kubenswrapper[4578]: I1003 13:37:24.752222 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-btppq" event={"ID":"3a3d676d-e014-4e02-b777-3a49b0ff11d1","Type":"ContainerStarted","Data":"46188c22fae41d036b0478e8fe1c23b2bba23b9f8ae6ff3bf7db535940cbd4ae"} Oct 03 13:37:25 crc kubenswrapper[4578]: I1003 13:37:25.763082 4578 generic.go:334] "Generic (PLEG): container finished" podID="e6ee1a77-57be-48c1-a61f-50719bfcbd29" containerID="aa66394aa10fbec4b87b93e4bd4f38c51b9f306e854210814f87bf571e79496d" exitCode=2 Oct 03 13:37:25 crc kubenswrapper[4578]: I1003 13:37:25.763157 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" event={"ID":"e6ee1a77-57be-48c1-a61f-50719bfcbd29","Type":"ContainerDied","Data":"aa66394aa10fbec4b87b93e4bd4f38c51b9f306e854210814f87bf571e79496d"} Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.188011 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.208064 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6ee1a77-57be-48c1-a61f-50719bfcbd29-ovn-combined-ca-bundle\") pod \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\" (UID: \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\") " Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.208104 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e6ee1a77-57be-48c1-a61f-50719bfcbd29-ovncontroller-config-0\") pod \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\" (UID: \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\") " Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.208179 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6ee1a77-57be-48c1-a61f-50719bfcbd29-ssh-key\") pod \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\" (UID: \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\") " Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.208202 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6ee1a77-57be-48c1-a61f-50719bfcbd29-inventory\") pod \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\" (UID: \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\") " Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.208227 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c569t\" (UniqueName: \"kubernetes.io/projected/e6ee1a77-57be-48c1-a61f-50719bfcbd29-kube-api-access-c569t\") pod \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\" (UID: \"e6ee1a77-57be-48c1-a61f-50719bfcbd29\") " Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.214537 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6ee1a77-57be-48c1-a61f-50719bfcbd29-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "e6ee1a77-57be-48c1-a61f-50719bfcbd29" (UID: "e6ee1a77-57be-48c1-a61f-50719bfcbd29"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.228847 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6ee1a77-57be-48c1-a61f-50719bfcbd29-kube-api-access-c569t" (OuterVolumeSpecName: "kube-api-access-c569t") pod "e6ee1a77-57be-48c1-a61f-50719bfcbd29" (UID: "e6ee1a77-57be-48c1-a61f-50719bfcbd29"). InnerVolumeSpecName "kube-api-access-c569t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.248431 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6ee1a77-57be-48c1-a61f-50719bfcbd29-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e6ee1a77-57be-48c1-a61f-50719bfcbd29" (UID: "e6ee1a77-57be-48c1-a61f-50719bfcbd29"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.255309 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6ee1a77-57be-48c1-a61f-50719bfcbd29-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "e6ee1a77-57be-48c1-a61f-50719bfcbd29" (UID: "e6ee1a77-57be-48c1-a61f-50719bfcbd29"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:37:27 crc kubenswrapper[4578]: E1003 13:37:27.256042 4578 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a3d676d_e014_4e02_b777_3a49b0ff11d1.slice/crio-conmon-46188c22fae41d036b0478e8fe1c23b2bba23b9f8ae6ff3bf7db535940cbd4ae.scope\": RecentStats: unable to find data in memory cache]" Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.272245 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6ee1a77-57be-48c1-a61f-50719bfcbd29-inventory" (OuterVolumeSpecName: "inventory") pod "e6ee1a77-57be-48c1-a61f-50719bfcbd29" (UID: "e6ee1a77-57be-48c1-a61f-50719bfcbd29"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.312178 4578 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6ee1a77-57be-48c1-a61f-50719bfcbd29-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.312832 4578 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e6ee1a77-57be-48c1-a61f-50719bfcbd29-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.312969 4578 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6ee1a77-57be-48c1-a61f-50719bfcbd29-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.313043 4578 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6ee1a77-57be-48c1-a61f-50719bfcbd29-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.313122 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c569t\" (UniqueName: \"kubernetes.io/projected/e6ee1a77-57be-48c1-a61f-50719bfcbd29-kube-api-access-c569t\") on node \"crc\" DevicePath \"\"" Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.781669 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" event={"ID":"e6ee1a77-57be-48c1-a61f-50719bfcbd29","Type":"ContainerDied","Data":"9b8dc242ae34d52b502c0672df796e35773948cec75a30e8c162b1cbe1674e3b"} Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.781715 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b8dc242ae34d52b502c0672df796e35773948cec75a30e8c162b1cbe1674e3b" Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.781680 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7ztl6" Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.786887 4578 generic.go:334] "Generic (PLEG): container finished" podID="3a3d676d-e014-4e02-b777-3a49b0ff11d1" containerID="46188c22fae41d036b0478e8fe1c23b2bba23b9f8ae6ff3bf7db535940cbd4ae" exitCode=0 Oct 03 13:37:27 crc kubenswrapper[4578]: I1003 13:37:27.786937 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-btppq" event={"ID":"3a3d676d-e014-4e02-b777-3a49b0ff11d1","Type":"ContainerDied","Data":"46188c22fae41d036b0478e8fe1c23b2bba23b9f8ae6ff3bf7db535940cbd4ae"} Oct 03 13:37:28 crc kubenswrapper[4578]: I1003 13:37:28.801762 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-btppq" event={"ID":"3a3d676d-e014-4e02-b777-3a49b0ff11d1","Type":"ContainerStarted","Data":"fb8ebd99b00f6784870e1c1713bde246dac0892e668c2b36da932802235d4366"} Oct 03 13:37:28 crc kubenswrapper[4578]: I1003 13:37:28.829071 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-btppq" podStartSLOduration=2.144437661 podStartE2EDuration="6.829055854s" podCreationTimestamp="2025-10-03 13:37:22 +0000 UTC" firstStartedPulling="2025-10-03 13:37:23.744117782 +0000 UTC m=+2779.542589966" lastFinishedPulling="2025-10-03 13:37:28.428735975 +0000 UTC m=+2784.227208159" observedRunningTime="2025-10-03 13:37:28.827993821 +0000 UTC m=+2784.626466015" watchObservedRunningTime="2025-10-03 13:37:28.829055854 +0000 UTC m=+2784.627528028" Oct 03 13:37:32 crc kubenswrapper[4578]: I1003 13:37:32.451989 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-btppq" Oct 03 13:37:32 crc kubenswrapper[4578]: I1003 13:37:32.452705 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-btppq" Oct 03 13:37:32 crc kubenswrapper[4578]: I1003 13:37:32.505026 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-btppq" Oct 03 13:37:42 crc kubenswrapper[4578]: I1003 13:37:42.495564 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-btppq" Oct 03 13:37:42 crc kubenswrapper[4578]: I1003 13:37:42.540098 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-btppq"] Oct 03 13:37:42 crc kubenswrapper[4578]: I1003 13:37:42.945368 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-btppq" podUID="3a3d676d-e014-4e02-b777-3a49b0ff11d1" containerName="registry-server" containerID="cri-o://fb8ebd99b00f6784870e1c1713bde246dac0892e668c2b36da932802235d4366" gracePeriod=2 Oct 03 13:37:43 crc kubenswrapper[4578]: I1003 13:37:43.400369 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-btppq" Oct 03 13:37:43 crc kubenswrapper[4578]: I1003 13:37:43.540981 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a3d676d-e014-4e02-b777-3a49b0ff11d1-utilities\") pod \"3a3d676d-e014-4e02-b777-3a49b0ff11d1\" (UID: \"3a3d676d-e014-4e02-b777-3a49b0ff11d1\") " Oct 03 13:37:43 crc kubenswrapper[4578]: I1003 13:37:43.541188 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmn8p\" (UniqueName: \"kubernetes.io/projected/3a3d676d-e014-4e02-b777-3a49b0ff11d1-kube-api-access-mmn8p\") pod \"3a3d676d-e014-4e02-b777-3a49b0ff11d1\" (UID: \"3a3d676d-e014-4e02-b777-3a49b0ff11d1\") " Oct 03 13:37:43 crc kubenswrapper[4578]: I1003 13:37:43.541279 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a3d676d-e014-4e02-b777-3a49b0ff11d1-catalog-content\") pod \"3a3d676d-e014-4e02-b777-3a49b0ff11d1\" (UID: \"3a3d676d-e014-4e02-b777-3a49b0ff11d1\") " Oct 03 13:37:43 crc kubenswrapper[4578]: I1003 13:37:43.542227 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a3d676d-e014-4e02-b777-3a49b0ff11d1-utilities" (OuterVolumeSpecName: "utilities") pod "3a3d676d-e014-4e02-b777-3a49b0ff11d1" (UID: "3a3d676d-e014-4e02-b777-3a49b0ff11d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:37:43 crc kubenswrapper[4578]: I1003 13:37:43.548188 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a3d676d-e014-4e02-b777-3a49b0ff11d1-kube-api-access-mmn8p" (OuterVolumeSpecName: "kube-api-access-mmn8p") pod "3a3d676d-e014-4e02-b777-3a49b0ff11d1" (UID: "3a3d676d-e014-4e02-b777-3a49b0ff11d1"). InnerVolumeSpecName "kube-api-access-mmn8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:37:43 crc kubenswrapper[4578]: I1003 13:37:43.590992 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a3d676d-e014-4e02-b777-3a49b0ff11d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a3d676d-e014-4e02-b777-3a49b0ff11d1" (UID: "3a3d676d-e014-4e02-b777-3a49b0ff11d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:37:43 crc kubenswrapper[4578]: I1003 13:37:43.644408 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a3d676d-e014-4e02-b777-3a49b0ff11d1-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:37:43 crc kubenswrapper[4578]: I1003 13:37:43.644450 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmn8p\" (UniqueName: \"kubernetes.io/projected/3a3d676d-e014-4e02-b777-3a49b0ff11d1-kube-api-access-mmn8p\") on node \"crc\" DevicePath \"\"" Oct 03 13:37:43 crc kubenswrapper[4578]: I1003 13:37:43.644460 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a3d676d-e014-4e02-b777-3a49b0ff11d1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:37:43 crc kubenswrapper[4578]: I1003 13:37:43.959034 4578 generic.go:334] "Generic (PLEG): container finished" podID="3a3d676d-e014-4e02-b777-3a49b0ff11d1" containerID="fb8ebd99b00f6784870e1c1713bde246dac0892e668c2b36da932802235d4366" exitCode=0 Oct 03 13:37:43 crc kubenswrapper[4578]: I1003 13:37:43.959072 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-btppq" event={"ID":"3a3d676d-e014-4e02-b777-3a49b0ff11d1","Type":"ContainerDied","Data":"fb8ebd99b00f6784870e1c1713bde246dac0892e668c2b36da932802235d4366"} Oct 03 13:37:43 crc kubenswrapper[4578]: I1003 13:37:43.959098 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-btppq" event={"ID":"3a3d676d-e014-4e02-b777-3a49b0ff11d1","Type":"ContainerDied","Data":"a69eb8075b58731f2e19a581fc157cd1ee2366a36ddb2b98b1ea81661594adb9"} Oct 03 13:37:43 crc kubenswrapper[4578]: I1003 13:37:43.959117 4578 scope.go:117] "RemoveContainer" containerID="fb8ebd99b00f6784870e1c1713bde246dac0892e668c2b36da932802235d4366" Oct 03 13:37:43 crc kubenswrapper[4578]: I1003 13:37:43.959208 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-btppq" Oct 03 13:37:43 crc kubenswrapper[4578]: I1003 13:37:43.983385 4578 scope.go:117] "RemoveContainer" containerID="46188c22fae41d036b0478e8fe1c23b2bba23b9f8ae6ff3bf7db535940cbd4ae" Oct 03 13:37:44 crc kubenswrapper[4578]: I1003 13:37:44.010704 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-btppq"] Oct 03 13:37:44 crc kubenswrapper[4578]: I1003 13:37:44.018620 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-btppq"] Oct 03 13:37:44 crc kubenswrapper[4578]: I1003 13:37:44.020055 4578 scope.go:117] "RemoveContainer" containerID="8ed1f421bbeb8edb6464024799a5b45584846ce8690c812ad2c4404ac690b8ee" Oct 03 13:37:44 crc kubenswrapper[4578]: I1003 13:37:44.051569 4578 scope.go:117] "RemoveContainer" containerID="fb8ebd99b00f6784870e1c1713bde246dac0892e668c2b36da932802235d4366" Oct 03 13:37:44 crc kubenswrapper[4578]: E1003 13:37:44.051959 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb8ebd99b00f6784870e1c1713bde246dac0892e668c2b36da932802235d4366\": container with ID starting with fb8ebd99b00f6784870e1c1713bde246dac0892e668c2b36da932802235d4366 not found: ID does not exist" containerID="fb8ebd99b00f6784870e1c1713bde246dac0892e668c2b36da932802235d4366" Oct 03 13:37:44 crc kubenswrapper[4578]: I1003 13:37:44.051989 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb8ebd99b00f6784870e1c1713bde246dac0892e668c2b36da932802235d4366"} err="failed to get container status \"fb8ebd99b00f6784870e1c1713bde246dac0892e668c2b36da932802235d4366\": rpc error: code = NotFound desc = could not find container \"fb8ebd99b00f6784870e1c1713bde246dac0892e668c2b36da932802235d4366\": container with ID starting with fb8ebd99b00f6784870e1c1713bde246dac0892e668c2b36da932802235d4366 not found: ID does not exist" Oct 03 13:37:44 crc kubenswrapper[4578]: I1003 13:37:44.052007 4578 scope.go:117] "RemoveContainer" containerID="46188c22fae41d036b0478e8fe1c23b2bba23b9f8ae6ff3bf7db535940cbd4ae" Oct 03 13:37:44 crc kubenswrapper[4578]: E1003 13:37:44.052314 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46188c22fae41d036b0478e8fe1c23b2bba23b9f8ae6ff3bf7db535940cbd4ae\": container with ID starting with 46188c22fae41d036b0478e8fe1c23b2bba23b9f8ae6ff3bf7db535940cbd4ae not found: ID does not exist" containerID="46188c22fae41d036b0478e8fe1c23b2bba23b9f8ae6ff3bf7db535940cbd4ae" Oct 03 13:37:44 crc kubenswrapper[4578]: I1003 13:37:44.052336 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46188c22fae41d036b0478e8fe1c23b2bba23b9f8ae6ff3bf7db535940cbd4ae"} err="failed to get container status \"46188c22fae41d036b0478e8fe1c23b2bba23b9f8ae6ff3bf7db535940cbd4ae\": rpc error: code = NotFound desc = could not find container \"46188c22fae41d036b0478e8fe1c23b2bba23b9f8ae6ff3bf7db535940cbd4ae\": container with ID starting with 46188c22fae41d036b0478e8fe1c23b2bba23b9f8ae6ff3bf7db535940cbd4ae not found: ID does not exist" Oct 03 13:37:44 crc kubenswrapper[4578]: I1003 13:37:44.052349 4578 scope.go:117] "RemoveContainer" containerID="8ed1f421bbeb8edb6464024799a5b45584846ce8690c812ad2c4404ac690b8ee" Oct 03 13:37:44 crc kubenswrapper[4578]: E1003 13:37:44.052570 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ed1f421bbeb8edb6464024799a5b45584846ce8690c812ad2c4404ac690b8ee\": container with ID starting with 8ed1f421bbeb8edb6464024799a5b45584846ce8690c812ad2c4404ac690b8ee not found: ID does not exist" containerID="8ed1f421bbeb8edb6464024799a5b45584846ce8690c812ad2c4404ac690b8ee" Oct 03 13:37:44 crc kubenswrapper[4578]: I1003 13:37:44.052589 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ed1f421bbeb8edb6464024799a5b45584846ce8690c812ad2c4404ac690b8ee"} err="failed to get container status \"8ed1f421bbeb8edb6464024799a5b45584846ce8690c812ad2c4404ac690b8ee\": rpc error: code = NotFound desc = could not find container \"8ed1f421bbeb8edb6464024799a5b45584846ce8690c812ad2c4404ac690b8ee\": container with ID starting with 8ed1f421bbeb8edb6464024799a5b45584846ce8690c812ad2c4404ac690b8ee not found: ID does not exist" Oct 03 13:37:44 crc kubenswrapper[4578]: I1003 13:37:44.919877 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a3d676d-e014-4e02-b777-3a49b0ff11d1" path="/var/lib/kubelet/pods/3a3d676d-e014-4e02-b777-3a49b0ff11d1/volumes" Oct 03 13:38:05 crc kubenswrapper[4578]: I1003 13:38:05.091563 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:38:05 crc kubenswrapper[4578]: I1003 13:38:05.092228 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:38:35 crc kubenswrapper[4578]: I1003 13:38:35.091610 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:38:35 crc kubenswrapper[4578]: I1003 13:38:35.092501 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:39:05 crc kubenswrapper[4578]: I1003 13:39:05.091456 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:39:05 crc kubenswrapper[4578]: I1003 13:39:05.091965 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:39:05 crc kubenswrapper[4578]: I1003 13:39:05.092008 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 13:39:05 crc kubenswrapper[4578]: I1003 13:39:05.092710 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406"} pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:39:05 crc kubenswrapper[4578]: I1003 13:39:05.092765 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" containerID="cri-o://f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" gracePeriod=600 Oct 03 13:39:05 crc kubenswrapper[4578]: E1003 13:39:05.227776 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:39:05 crc kubenswrapper[4578]: I1003 13:39:05.715834 4578 generic.go:334] "Generic (PLEG): container finished" podID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" exitCode=0 Oct 03 13:39:05 crc kubenswrapper[4578]: I1003 13:39:05.715875 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerDied","Data":"f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406"} Oct 03 13:39:05 crc kubenswrapper[4578]: I1003 13:39:05.715906 4578 scope.go:117] "RemoveContainer" containerID="32a0b82351869c87347f7c4f14ff5c12356fd24347e53a70e720916381ba5d60" Oct 03 13:39:05 crc kubenswrapper[4578]: I1003 13:39:05.716760 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:39:05 crc kubenswrapper[4578]: E1003 13:39:05.717088 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:39:10 crc kubenswrapper[4578]: I1003 13:39:10.210232 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t7t7m"] Oct 03 13:39:10 crc kubenswrapper[4578]: E1003 13:39:10.211488 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a3d676d-e014-4e02-b777-3a49b0ff11d1" containerName="registry-server" Oct 03 13:39:10 crc kubenswrapper[4578]: I1003 13:39:10.211503 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a3d676d-e014-4e02-b777-3a49b0ff11d1" containerName="registry-server" Oct 03 13:39:10 crc kubenswrapper[4578]: E1003 13:39:10.211526 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a3d676d-e014-4e02-b777-3a49b0ff11d1" containerName="extract-content" Oct 03 13:39:10 crc kubenswrapper[4578]: I1003 13:39:10.211533 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a3d676d-e014-4e02-b777-3a49b0ff11d1" containerName="extract-content" Oct 03 13:39:10 crc kubenswrapper[4578]: E1003 13:39:10.211560 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6ee1a77-57be-48c1-a61f-50719bfcbd29" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:39:10 crc kubenswrapper[4578]: I1003 13:39:10.211624 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6ee1a77-57be-48c1-a61f-50719bfcbd29" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:39:10 crc kubenswrapper[4578]: E1003 13:39:10.211669 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a3d676d-e014-4e02-b777-3a49b0ff11d1" containerName="extract-utilities" Oct 03 13:39:10 crc kubenswrapper[4578]: I1003 13:39:10.211676 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a3d676d-e014-4e02-b777-3a49b0ff11d1" containerName="extract-utilities" Oct 03 13:39:10 crc kubenswrapper[4578]: I1003 13:39:10.212237 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a3d676d-e014-4e02-b777-3a49b0ff11d1" containerName="registry-server" Oct 03 13:39:10 crc kubenswrapper[4578]: I1003 13:39:10.212266 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6ee1a77-57be-48c1-a61f-50719bfcbd29" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:39:10 crc kubenswrapper[4578]: I1003 13:39:10.215044 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t7t7m" Oct 03 13:39:10 crc kubenswrapper[4578]: I1003 13:39:10.236161 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68375be5-cb25-4ffa-b3a7-cac00f7d6071-catalog-content\") pod \"redhat-operators-t7t7m\" (UID: \"68375be5-cb25-4ffa-b3a7-cac00f7d6071\") " pod="openshift-marketplace/redhat-operators-t7t7m" Oct 03 13:39:10 crc kubenswrapper[4578]: I1003 13:39:10.236403 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68375be5-cb25-4ffa-b3a7-cac00f7d6071-utilities\") pod \"redhat-operators-t7t7m\" (UID: \"68375be5-cb25-4ffa-b3a7-cac00f7d6071\") " pod="openshift-marketplace/redhat-operators-t7t7m" Oct 03 13:39:10 crc kubenswrapper[4578]: I1003 13:39:10.236542 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2sdt\" (UniqueName: \"kubernetes.io/projected/68375be5-cb25-4ffa-b3a7-cac00f7d6071-kube-api-access-s2sdt\") pod \"redhat-operators-t7t7m\" (UID: \"68375be5-cb25-4ffa-b3a7-cac00f7d6071\") " pod="openshift-marketplace/redhat-operators-t7t7m" Oct 03 13:39:10 crc kubenswrapper[4578]: I1003 13:39:10.247611 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t7t7m"] Oct 03 13:39:10 crc kubenswrapper[4578]: I1003 13:39:10.343704 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2sdt\" (UniqueName: \"kubernetes.io/projected/68375be5-cb25-4ffa-b3a7-cac00f7d6071-kube-api-access-s2sdt\") pod \"redhat-operators-t7t7m\" (UID: \"68375be5-cb25-4ffa-b3a7-cac00f7d6071\") " pod="openshift-marketplace/redhat-operators-t7t7m" Oct 03 13:39:10 crc kubenswrapper[4578]: I1003 13:39:10.343809 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68375be5-cb25-4ffa-b3a7-cac00f7d6071-catalog-content\") pod \"redhat-operators-t7t7m\" (UID: \"68375be5-cb25-4ffa-b3a7-cac00f7d6071\") " pod="openshift-marketplace/redhat-operators-t7t7m" Oct 03 13:39:10 crc kubenswrapper[4578]: I1003 13:39:10.343888 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68375be5-cb25-4ffa-b3a7-cac00f7d6071-utilities\") pod \"redhat-operators-t7t7m\" (UID: \"68375be5-cb25-4ffa-b3a7-cac00f7d6071\") " pod="openshift-marketplace/redhat-operators-t7t7m" Oct 03 13:39:10 crc kubenswrapper[4578]: I1003 13:39:10.344359 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68375be5-cb25-4ffa-b3a7-cac00f7d6071-utilities\") pod \"redhat-operators-t7t7m\" (UID: \"68375be5-cb25-4ffa-b3a7-cac00f7d6071\") " pod="openshift-marketplace/redhat-operators-t7t7m" Oct 03 13:39:10 crc kubenswrapper[4578]: I1003 13:39:10.344439 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68375be5-cb25-4ffa-b3a7-cac00f7d6071-catalog-content\") pod \"redhat-operators-t7t7m\" (UID: \"68375be5-cb25-4ffa-b3a7-cac00f7d6071\") " pod="openshift-marketplace/redhat-operators-t7t7m" Oct 03 13:39:10 crc kubenswrapper[4578]: I1003 13:39:10.373045 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2sdt\" (UniqueName: \"kubernetes.io/projected/68375be5-cb25-4ffa-b3a7-cac00f7d6071-kube-api-access-s2sdt\") pod \"redhat-operators-t7t7m\" (UID: \"68375be5-cb25-4ffa-b3a7-cac00f7d6071\") " pod="openshift-marketplace/redhat-operators-t7t7m" Oct 03 13:39:10 crc kubenswrapper[4578]: I1003 13:39:10.549526 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t7t7m" Oct 03 13:39:11 crc kubenswrapper[4578]: I1003 13:39:11.015070 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t7t7m"] Oct 03 13:39:11 crc kubenswrapper[4578]: I1003 13:39:11.783375 4578 generic.go:334] "Generic (PLEG): container finished" podID="68375be5-cb25-4ffa-b3a7-cac00f7d6071" containerID="2d5a4e150ab18f85928d82b115ca5ba83d29a4db6b2a920b35aa77b3216dbd9d" exitCode=0 Oct 03 13:39:11 crc kubenswrapper[4578]: I1003 13:39:11.783523 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7t7m" event={"ID":"68375be5-cb25-4ffa-b3a7-cac00f7d6071","Type":"ContainerDied","Data":"2d5a4e150ab18f85928d82b115ca5ba83d29a4db6b2a920b35aa77b3216dbd9d"} Oct 03 13:39:11 crc kubenswrapper[4578]: I1003 13:39:11.783787 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7t7m" event={"ID":"68375be5-cb25-4ffa-b3a7-cac00f7d6071","Type":"ContainerStarted","Data":"5ee4e8b55422e6cf7ccfa4a12efd02c732e61c6b2fd7a5add032a9072202be78"} Oct 03 13:39:13 crc kubenswrapper[4578]: I1003 13:39:13.803267 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7t7m" event={"ID":"68375be5-cb25-4ffa-b3a7-cac00f7d6071","Type":"ContainerStarted","Data":"6651c7d9f78861d23fe5d9371c068949b5dbb549f00ef86484ae37c23f935801"} Oct 03 13:39:18 crc kubenswrapper[4578]: I1003 13:39:18.908969 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:39:18 crc kubenswrapper[4578]: E1003 13:39:18.910949 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:39:19 crc kubenswrapper[4578]: I1003 13:39:19.862174 4578 generic.go:334] "Generic (PLEG): container finished" podID="68375be5-cb25-4ffa-b3a7-cac00f7d6071" containerID="6651c7d9f78861d23fe5d9371c068949b5dbb549f00ef86484ae37c23f935801" exitCode=0 Oct 03 13:39:19 crc kubenswrapper[4578]: I1003 13:39:19.862216 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7t7m" event={"ID":"68375be5-cb25-4ffa-b3a7-cac00f7d6071","Type":"ContainerDied","Data":"6651c7d9f78861d23fe5d9371c068949b5dbb549f00ef86484ae37c23f935801"} Oct 03 13:39:21 crc kubenswrapper[4578]: I1003 13:39:21.885891 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7t7m" event={"ID":"68375be5-cb25-4ffa-b3a7-cac00f7d6071","Type":"ContainerStarted","Data":"9ce325f211fc6b44ab5d552fe697e44d56af1c906d28575227765921d702c0bb"} Oct 03 13:39:21 crc kubenswrapper[4578]: I1003 13:39:21.911298 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t7t7m" podStartSLOduration=3.024246484 podStartE2EDuration="11.911278331s" podCreationTimestamp="2025-10-03 13:39:10 +0000 UTC" firstStartedPulling="2025-10-03 13:39:11.786358497 +0000 UTC m=+2887.584830681" lastFinishedPulling="2025-10-03 13:39:20.673390344 +0000 UTC m=+2896.471862528" observedRunningTime="2025-10-03 13:39:21.906932006 +0000 UTC m=+2897.705404190" watchObservedRunningTime="2025-10-03 13:39:21.911278331 +0000 UTC m=+2897.709750515" Oct 03 13:39:30 crc kubenswrapper[4578]: I1003 13:39:30.550642 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t7t7m" Oct 03 13:39:30 crc kubenswrapper[4578]: I1003 13:39:30.551128 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t7t7m" Oct 03 13:39:30 crc kubenswrapper[4578]: I1003 13:39:30.597955 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t7t7m" Oct 03 13:39:31 crc kubenswrapper[4578]: I1003 13:39:31.019296 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t7t7m" Oct 03 13:39:31 crc kubenswrapper[4578]: I1003 13:39:31.067725 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t7t7m"] Oct 03 13:39:32 crc kubenswrapper[4578]: I1003 13:39:32.909819 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:39:32 crc kubenswrapper[4578]: E1003 13:39:32.910797 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:39:32 crc kubenswrapper[4578]: I1003 13:39:32.978363 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t7t7m" podUID="68375be5-cb25-4ffa-b3a7-cac00f7d6071" containerName="registry-server" containerID="cri-o://9ce325f211fc6b44ab5d552fe697e44d56af1c906d28575227765921d702c0bb" gracePeriod=2 Oct 03 13:39:33 crc kubenswrapper[4578]: I1003 13:39:33.434228 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t7t7m" Oct 03 13:39:33 crc kubenswrapper[4578]: I1003 13:39:33.560680 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68375be5-cb25-4ffa-b3a7-cac00f7d6071-catalog-content\") pod \"68375be5-cb25-4ffa-b3a7-cac00f7d6071\" (UID: \"68375be5-cb25-4ffa-b3a7-cac00f7d6071\") " Oct 03 13:39:33 crc kubenswrapper[4578]: I1003 13:39:33.560961 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68375be5-cb25-4ffa-b3a7-cac00f7d6071-utilities\") pod \"68375be5-cb25-4ffa-b3a7-cac00f7d6071\" (UID: \"68375be5-cb25-4ffa-b3a7-cac00f7d6071\") " Oct 03 13:39:33 crc kubenswrapper[4578]: I1003 13:39:33.561073 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2sdt\" (UniqueName: \"kubernetes.io/projected/68375be5-cb25-4ffa-b3a7-cac00f7d6071-kube-api-access-s2sdt\") pod \"68375be5-cb25-4ffa-b3a7-cac00f7d6071\" (UID: \"68375be5-cb25-4ffa-b3a7-cac00f7d6071\") " Oct 03 13:39:33 crc kubenswrapper[4578]: I1003 13:39:33.561987 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68375be5-cb25-4ffa-b3a7-cac00f7d6071-utilities" (OuterVolumeSpecName: "utilities") pod "68375be5-cb25-4ffa-b3a7-cac00f7d6071" (UID: "68375be5-cb25-4ffa-b3a7-cac00f7d6071"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:39:33 crc kubenswrapper[4578]: I1003 13:39:33.568990 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68375be5-cb25-4ffa-b3a7-cac00f7d6071-kube-api-access-s2sdt" (OuterVolumeSpecName: "kube-api-access-s2sdt") pod "68375be5-cb25-4ffa-b3a7-cac00f7d6071" (UID: "68375be5-cb25-4ffa-b3a7-cac00f7d6071"). InnerVolumeSpecName "kube-api-access-s2sdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:39:33 crc kubenswrapper[4578]: I1003 13:39:33.638672 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68375be5-cb25-4ffa-b3a7-cac00f7d6071-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68375be5-cb25-4ffa-b3a7-cac00f7d6071" (UID: "68375be5-cb25-4ffa-b3a7-cac00f7d6071"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:39:33 crc kubenswrapper[4578]: I1003 13:39:33.663199 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68375be5-cb25-4ffa-b3a7-cac00f7d6071-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:39:33 crc kubenswrapper[4578]: I1003 13:39:33.663460 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2sdt\" (UniqueName: \"kubernetes.io/projected/68375be5-cb25-4ffa-b3a7-cac00f7d6071-kube-api-access-s2sdt\") on node \"crc\" DevicePath \"\"" Oct 03 13:39:33 crc kubenswrapper[4578]: I1003 13:39:33.663529 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68375be5-cb25-4ffa-b3a7-cac00f7d6071-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:39:33 crc kubenswrapper[4578]: I1003 13:39:33.989120 4578 generic.go:334] "Generic (PLEG): container finished" podID="68375be5-cb25-4ffa-b3a7-cac00f7d6071" containerID="9ce325f211fc6b44ab5d552fe697e44d56af1c906d28575227765921d702c0bb" exitCode=0 Oct 03 13:39:33 crc kubenswrapper[4578]: I1003 13:39:33.989159 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7t7m" event={"ID":"68375be5-cb25-4ffa-b3a7-cac00f7d6071","Type":"ContainerDied","Data":"9ce325f211fc6b44ab5d552fe697e44d56af1c906d28575227765921d702c0bb"} Oct 03 13:39:33 crc kubenswrapper[4578]: I1003 13:39:33.989169 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t7t7m" Oct 03 13:39:33 crc kubenswrapper[4578]: I1003 13:39:33.989183 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7t7m" event={"ID":"68375be5-cb25-4ffa-b3a7-cac00f7d6071","Type":"ContainerDied","Data":"5ee4e8b55422e6cf7ccfa4a12efd02c732e61c6b2fd7a5add032a9072202be78"} Oct 03 13:39:33 crc kubenswrapper[4578]: I1003 13:39:33.989200 4578 scope.go:117] "RemoveContainer" containerID="9ce325f211fc6b44ab5d552fe697e44d56af1c906d28575227765921d702c0bb" Oct 03 13:39:34 crc kubenswrapper[4578]: I1003 13:39:34.014502 4578 scope.go:117] "RemoveContainer" containerID="6651c7d9f78861d23fe5d9371c068949b5dbb549f00ef86484ae37c23f935801" Oct 03 13:39:34 crc kubenswrapper[4578]: I1003 13:39:34.028784 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t7t7m"] Oct 03 13:39:34 crc kubenswrapper[4578]: I1003 13:39:34.037566 4578 scope.go:117] "RemoveContainer" containerID="2d5a4e150ab18f85928d82b115ca5ba83d29a4db6b2a920b35aa77b3216dbd9d" Oct 03 13:39:34 crc kubenswrapper[4578]: I1003 13:39:34.040284 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t7t7m"] Oct 03 13:39:34 crc kubenswrapper[4578]: I1003 13:39:34.081336 4578 scope.go:117] "RemoveContainer" containerID="9ce325f211fc6b44ab5d552fe697e44d56af1c906d28575227765921d702c0bb" Oct 03 13:39:34 crc kubenswrapper[4578]: E1003 13:39:34.081889 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ce325f211fc6b44ab5d552fe697e44d56af1c906d28575227765921d702c0bb\": container with ID starting with 9ce325f211fc6b44ab5d552fe697e44d56af1c906d28575227765921d702c0bb not found: ID does not exist" containerID="9ce325f211fc6b44ab5d552fe697e44d56af1c906d28575227765921d702c0bb" Oct 03 13:39:34 crc kubenswrapper[4578]: I1003 13:39:34.081927 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ce325f211fc6b44ab5d552fe697e44d56af1c906d28575227765921d702c0bb"} err="failed to get container status \"9ce325f211fc6b44ab5d552fe697e44d56af1c906d28575227765921d702c0bb\": rpc error: code = NotFound desc = could not find container \"9ce325f211fc6b44ab5d552fe697e44d56af1c906d28575227765921d702c0bb\": container with ID starting with 9ce325f211fc6b44ab5d552fe697e44d56af1c906d28575227765921d702c0bb not found: ID does not exist" Oct 03 13:39:34 crc kubenswrapper[4578]: I1003 13:39:34.081954 4578 scope.go:117] "RemoveContainer" containerID="6651c7d9f78861d23fe5d9371c068949b5dbb549f00ef86484ae37c23f935801" Oct 03 13:39:34 crc kubenswrapper[4578]: E1003 13:39:34.082233 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6651c7d9f78861d23fe5d9371c068949b5dbb549f00ef86484ae37c23f935801\": container with ID starting with 6651c7d9f78861d23fe5d9371c068949b5dbb549f00ef86484ae37c23f935801 not found: ID does not exist" containerID="6651c7d9f78861d23fe5d9371c068949b5dbb549f00ef86484ae37c23f935801" Oct 03 13:39:34 crc kubenswrapper[4578]: I1003 13:39:34.082261 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6651c7d9f78861d23fe5d9371c068949b5dbb549f00ef86484ae37c23f935801"} err="failed to get container status \"6651c7d9f78861d23fe5d9371c068949b5dbb549f00ef86484ae37c23f935801\": rpc error: code = NotFound desc = could not find container \"6651c7d9f78861d23fe5d9371c068949b5dbb549f00ef86484ae37c23f935801\": container with ID starting with 6651c7d9f78861d23fe5d9371c068949b5dbb549f00ef86484ae37c23f935801 not found: ID does not exist" Oct 03 13:39:34 crc kubenswrapper[4578]: I1003 13:39:34.082278 4578 scope.go:117] "RemoveContainer" containerID="2d5a4e150ab18f85928d82b115ca5ba83d29a4db6b2a920b35aa77b3216dbd9d" Oct 03 13:39:34 crc kubenswrapper[4578]: E1003 13:39:34.082486 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d5a4e150ab18f85928d82b115ca5ba83d29a4db6b2a920b35aa77b3216dbd9d\": container with ID starting with 2d5a4e150ab18f85928d82b115ca5ba83d29a4db6b2a920b35aa77b3216dbd9d not found: ID does not exist" containerID="2d5a4e150ab18f85928d82b115ca5ba83d29a4db6b2a920b35aa77b3216dbd9d" Oct 03 13:39:34 crc kubenswrapper[4578]: I1003 13:39:34.082514 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d5a4e150ab18f85928d82b115ca5ba83d29a4db6b2a920b35aa77b3216dbd9d"} err="failed to get container status \"2d5a4e150ab18f85928d82b115ca5ba83d29a4db6b2a920b35aa77b3216dbd9d\": rpc error: code = NotFound desc = could not find container \"2d5a4e150ab18f85928d82b115ca5ba83d29a4db6b2a920b35aa77b3216dbd9d\": container with ID starting with 2d5a4e150ab18f85928d82b115ca5ba83d29a4db6b2a920b35aa77b3216dbd9d not found: ID does not exist" Oct 03 13:39:34 crc kubenswrapper[4578]: I1003 13:39:34.924465 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68375be5-cb25-4ffa-b3a7-cac00f7d6071" path="/var/lib/kubelet/pods/68375be5-cb25-4ffa-b3a7-cac00f7d6071/volumes" Oct 03 13:39:45 crc kubenswrapper[4578]: I1003 13:39:45.909737 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:39:45 crc kubenswrapper[4578]: E1003 13:39:45.910810 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:39:57 crc kubenswrapper[4578]: I1003 13:39:57.909355 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:39:57 crc kubenswrapper[4578]: E1003 13:39:57.910105 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:40:11 crc kubenswrapper[4578]: I1003 13:40:11.909311 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:40:11 crc kubenswrapper[4578]: E1003 13:40:11.910046 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:40:26 crc kubenswrapper[4578]: I1003 13:40:26.909792 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:40:26 crc kubenswrapper[4578]: E1003 13:40:26.911396 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:40:39 crc kubenswrapper[4578]: I1003 13:40:39.909875 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:40:39 crc kubenswrapper[4578]: E1003 13:40:39.912047 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:40:50 crc kubenswrapper[4578]: I1003 13:40:50.909555 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:40:50 crc kubenswrapper[4578]: E1003 13:40:50.910430 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:41:02 crc kubenswrapper[4578]: I1003 13:41:02.909520 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:41:02 crc kubenswrapper[4578]: E1003 13:41:02.910360 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:41:13 crc kubenswrapper[4578]: I1003 13:41:13.908918 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:41:13 crc kubenswrapper[4578]: E1003 13:41:13.909768 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:41:16 crc kubenswrapper[4578]: I1003 13:41:16.813959 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lsbxs"] Oct 03 13:41:16 crc kubenswrapper[4578]: E1003 13:41:16.814726 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68375be5-cb25-4ffa-b3a7-cac00f7d6071" containerName="extract-content" Oct 03 13:41:16 crc kubenswrapper[4578]: I1003 13:41:16.814743 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="68375be5-cb25-4ffa-b3a7-cac00f7d6071" containerName="extract-content" Oct 03 13:41:16 crc kubenswrapper[4578]: E1003 13:41:16.814756 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68375be5-cb25-4ffa-b3a7-cac00f7d6071" containerName="extract-utilities" Oct 03 13:41:16 crc kubenswrapper[4578]: I1003 13:41:16.814763 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="68375be5-cb25-4ffa-b3a7-cac00f7d6071" containerName="extract-utilities" Oct 03 13:41:16 crc kubenswrapper[4578]: E1003 13:41:16.814794 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68375be5-cb25-4ffa-b3a7-cac00f7d6071" containerName="registry-server" Oct 03 13:41:16 crc kubenswrapper[4578]: I1003 13:41:16.814801 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="68375be5-cb25-4ffa-b3a7-cac00f7d6071" containerName="registry-server" Oct 03 13:41:16 crc kubenswrapper[4578]: I1003 13:41:16.815054 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="68375be5-cb25-4ffa-b3a7-cac00f7d6071" containerName="registry-server" Oct 03 13:41:16 crc kubenswrapper[4578]: I1003 13:41:16.817508 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lsbxs" Oct 03 13:41:16 crc kubenswrapper[4578]: I1003 13:41:16.831817 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lsbxs"] Oct 03 13:41:16 crc kubenswrapper[4578]: I1003 13:41:16.964751 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72c39a37-24c7-4e53-8f1b-b3a1794b9053-utilities\") pod \"redhat-marketplace-lsbxs\" (UID: \"72c39a37-24c7-4e53-8f1b-b3a1794b9053\") " pod="openshift-marketplace/redhat-marketplace-lsbxs" Oct 03 13:41:16 crc kubenswrapper[4578]: I1003 13:41:16.964833 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qqkz\" (UniqueName: \"kubernetes.io/projected/72c39a37-24c7-4e53-8f1b-b3a1794b9053-kube-api-access-4qqkz\") pod \"redhat-marketplace-lsbxs\" (UID: \"72c39a37-24c7-4e53-8f1b-b3a1794b9053\") " pod="openshift-marketplace/redhat-marketplace-lsbxs" Oct 03 13:41:16 crc kubenswrapper[4578]: I1003 13:41:16.964905 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72c39a37-24c7-4e53-8f1b-b3a1794b9053-catalog-content\") pod \"redhat-marketplace-lsbxs\" (UID: \"72c39a37-24c7-4e53-8f1b-b3a1794b9053\") " pod="openshift-marketplace/redhat-marketplace-lsbxs" Oct 03 13:41:17 crc kubenswrapper[4578]: I1003 13:41:17.066571 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72c39a37-24c7-4e53-8f1b-b3a1794b9053-utilities\") pod \"redhat-marketplace-lsbxs\" (UID: \"72c39a37-24c7-4e53-8f1b-b3a1794b9053\") " pod="openshift-marketplace/redhat-marketplace-lsbxs" Oct 03 13:41:17 crc kubenswrapper[4578]: I1003 13:41:17.066658 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qqkz\" (UniqueName: \"kubernetes.io/projected/72c39a37-24c7-4e53-8f1b-b3a1794b9053-kube-api-access-4qqkz\") pod \"redhat-marketplace-lsbxs\" (UID: \"72c39a37-24c7-4e53-8f1b-b3a1794b9053\") " pod="openshift-marketplace/redhat-marketplace-lsbxs" Oct 03 13:41:17 crc kubenswrapper[4578]: I1003 13:41:17.066697 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72c39a37-24c7-4e53-8f1b-b3a1794b9053-catalog-content\") pod \"redhat-marketplace-lsbxs\" (UID: \"72c39a37-24c7-4e53-8f1b-b3a1794b9053\") " pod="openshift-marketplace/redhat-marketplace-lsbxs" Oct 03 13:41:17 crc kubenswrapper[4578]: I1003 13:41:17.067060 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72c39a37-24c7-4e53-8f1b-b3a1794b9053-utilities\") pod \"redhat-marketplace-lsbxs\" (UID: \"72c39a37-24c7-4e53-8f1b-b3a1794b9053\") " pod="openshift-marketplace/redhat-marketplace-lsbxs" Oct 03 13:41:17 crc kubenswrapper[4578]: I1003 13:41:17.067289 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72c39a37-24c7-4e53-8f1b-b3a1794b9053-catalog-content\") pod \"redhat-marketplace-lsbxs\" (UID: \"72c39a37-24c7-4e53-8f1b-b3a1794b9053\") " pod="openshift-marketplace/redhat-marketplace-lsbxs" Oct 03 13:41:17 crc kubenswrapper[4578]: I1003 13:41:17.088473 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qqkz\" (UniqueName: \"kubernetes.io/projected/72c39a37-24c7-4e53-8f1b-b3a1794b9053-kube-api-access-4qqkz\") pod \"redhat-marketplace-lsbxs\" (UID: \"72c39a37-24c7-4e53-8f1b-b3a1794b9053\") " pod="openshift-marketplace/redhat-marketplace-lsbxs" Oct 03 13:41:17 crc kubenswrapper[4578]: I1003 13:41:17.149022 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lsbxs" Oct 03 13:41:17 crc kubenswrapper[4578]: I1003 13:41:17.618328 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lsbxs"] Oct 03 13:41:17 crc kubenswrapper[4578]: I1003 13:41:17.899978 4578 generic.go:334] "Generic (PLEG): container finished" podID="72c39a37-24c7-4e53-8f1b-b3a1794b9053" containerID="5f69bdeb0d4e61f9fea44b24213b6a747136bab8e0ac077ad55457b446dcc8ac" exitCode=0 Oct 03 13:41:17 crc kubenswrapper[4578]: I1003 13:41:17.901327 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lsbxs" event={"ID":"72c39a37-24c7-4e53-8f1b-b3a1794b9053","Type":"ContainerDied","Data":"5f69bdeb0d4e61f9fea44b24213b6a747136bab8e0ac077ad55457b446dcc8ac"} Oct 03 13:41:17 crc kubenswrapper[4578]: I1003 13:41:17.901439 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lsbxs" event={"ID":"72c39a37-24c7-4e53-8f1b-b3a1794b9053","Type":"ContainerStarted","Data":"f5d403f36d265817c7daea43950d6721ca8a2260c7bff5005905a7df0a7ec16b"} Oct 03 13:41:19 crc kubenswrapper[4578]: I1003 13:41:19.923568 4578 generic.go:334] "Generic (PLEG): container finished" podID="72c39a37-24c7-4e53-8f1b-b3a1794b9053" containerID="b9d656a3393eb5754f58714370a84898b144266c932eb399cd45b16ce7a45fa4" exitCode=0 Oct 03 13:41:19 crc kubenswrapper[4578]: I1003 13:41:19.923662 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lsbxs" event={"ID":"72c39a37-24c7-4e53-8f1b-b3a1794b9053","Type":"ContainerDied","Data":"b9d656a3393eb5754f58714370a84898b144266c932eb399cd45b16ce7a45fa4"} Oct 03 13:41:20 crc kubenswrapper[4578]: I1003 13:41:20.939038 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lsbxs" event={"ID":"72c39a37-24c7-4e53-8f1b-b3a1794b9053","Type":"ContainerStarted","Data":"e2533197f40eea8fb13381b8c5d746cdef66e3db2066e291bcf6225eb969a6ea"} Oct 03 13:41:24 crc kubenswrapper[4578]: I1003 13:41:24.916307 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:41:24 crc kubenswrapper[4578]: E1003 13:41:24.917543 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:41:27 crc kubenswrapper[4578]: I1003 13:41:27.150460 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lsbxs" Oct 03 13:41:27 crc kubenswrapper[4578]: I1003 13:41:27.150814 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lsbxs" Oct 03 13:41:27 crc kubenswrapper[4578]: I1003 13:41:27.199982 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lsbxs" Oct 03 13:41:27 crc kubenswrapper[4578]: I1003 13:41:27.217693 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lsbxs" podStartSLOduration=8.672795348 podStartE2EDuration="11.217674103s" podCreationTimestamp="2025-10-03 13:41:16 +0000 UTC" firstStartedPulling="2025-10-03 13:41:17.903555438 +0000 UTC m=+3013.702027622" lastFinishedPulling="2025-10-03 13:41:20.448434193 +0000 UTC m=+3016.246906377" observedRunningTime="2025-10-03 13:41:20.960585687 +0000 UTC m=+3016.759057881" watchObservedRunningTime="2025-10-03 13:41:27.217674103 +0000 UTC m=+3023.016146287" Oct 03 13:41:28 crc kubenswrapper[4578]: I1003 13:41:28.047441 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lsbxs" Oct 03 13:41:28 crc kubenswrapper[4578]: I1003 13:41:28.101511 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lsbxs"] Oct 03 13:41:30 crc kubenswrapper[4578]: I1003 13:41:30.018187 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lsbxs" podUID="72c39a37-24c7-4e53-8f1b-b3a1794b9053" containerName="registry-server" containerID="cri-o://e2533197f40eea8fb13381b8c5d746cdef66e3db2066e291bcf6225eb969a6ea" gracePeriod=2 Oct 03 13:41:30 crc kubenswrapper[4578]: I1003 13:41:30.447678 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lsbxs" Oct 03 13:41:30 crc kubenswrapper[4578]: I1003 13:41:30.517075 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qqkz\" (UniqueName: \"kubernetes.io/projected/72c39a37-24c7-4e53-8f1b-b3a1794b9053-kube-api-access-4qqkz\") pod \"72c39a37-24c7-4e53-8f1b-b3a1794b9053\" (UID: \"72c39a37-24c7-4e53-8f1b-b3a1794b9053\") " Oct 03 13:41:30 crc kubenswrapper[4578]: I1003 13:41:30.517405 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72c39a37-24c7-4e53-8f1b-b3a1794b9053-utilities\") pod \"72c39a37-24c7-4e53-8f1b-b3a1794b9053\" (UID: \"72c39a37-24c7-4e53-8f1b-b3a1794b9053\") " Oct 03 13:41:30 crc kubenswrapper[4578]: I1003 13:41:30.517569 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72c39a37-24c7-4e53-8f1b-b3a1794b9053-catalog-content\") pod \"72c39a37-24c7-4e53-8f1b-b3a1794b9053\" (UID: \"72c39a37-24c7-4e53-8f1b-b3a1794b9053\") " Oct 03 13:41:30 crc kubenswrapper[4578]: I1003 13:41:30.518515 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72c39a37-24c7-4e53-8f1b-b3a1794b9053-utilities" (OuterVolumeSpecName: "utilities") pod "72c39a37-24c7-4e53-8f1b-b3a1794b9053" (UID: "72c39a37-24c7-4e53-8f1b-b3a1794b9053"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:41:30 crc kubenswrapper[4578]: I1003 13:41:30.524081 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72c39a37-24c7-4e53-8f1b-b3a1794b9053-kube-api-access-4qqkz" (OuterVolumeSpecName: "kube-api-access-4qqkz") pod "72c39a37-24c7-4e53-8f1b-b3a1794b9053" (UID: "72c39a37-24c7-4e53-8f1b-b3a1794b9053"). InnerVolumeSpecName "kube-api-access-4qqkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:41:30 crc kubenswrapper[4578]: I1003 13:41:30.533650 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72c39a37-24c7-4e53-8f1b-b3a1794b9053-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72c39a37-24c7-4e53-8f1b-b3a1794b9053" (UID: "72c39a37-24c7-4e53-8f1b-b3a1794b9053"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:41:30 crc kubenswrapper[4578]: I1003 13:41:30.621377 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qqkz\" (UniqueName: \"kubernetes.io/projected/72c39a37-24c7-4e53-8f1b-b3a1794b9053-kube-api-access-4qqkz\") on node \"crc\" DevicePath \"\"" Oct 03 13:41:30 crc kubenswrapper[4578]: I1003 13:41:30.621423 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72c39a37-24c7-4e53-8f1b-b3a1794b9053-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:41:30 crc kubenswrapper[4578]: I1003 13:41:30.621435 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72c39a37-24c7-4e53-8f1b-b3a1794b9053-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:41:31 crc kubenswrapper[4578]: I1003 13:41:31.028666 4578 generic.go:334] "Generic (PLEG): container finished" podID="72c39a37-24c7-4e53-8f1b-b3a1794b9053" containerID="e2533197f40eea8fb13381b8c5d746cdef66e3db2066e291bcf6225eb969a6ea" exitCode=0 Oct 03 13:41:31 crc kubenswrapper[4578]: I1003 13:41:31.028773 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lsbxs" Oct 03 13:41:31 crc kubenswrapper[4578]: I1003 13:41:31.028782 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lsbxs" event={"ID":"72c39a37-24c7-4e53-8f1b-b3a1794b9053","Type":"ContainerDied","Data":"e2533197f40eea8fb13381b8c5d746cdef66e3db2066e291bcf6225eb969a6ea"} Oct 03 13:41:31 crc kubenswrapper[4578]: I1003 13:41:31.028879 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lsbxs" event={"ID":"72c39a37-24c7-4e53-8f1b-b3a1794b9053","Type":"ContainerDied","Data":"f5d403f36d265817c7daea43950d6721ca8a2260c7bff5005905a7df0a7ec16b"} Oct 03 13:41:31 crc kubenswrapper[4578]: I1003 13:41:31.028905 4578 scope.go:117] "RemoveContainer" containerID="e2533197f40eea8fb13381b8c5d746cdef66e3db2066e291bcf6225eb969a6ea" Oct 03 13:41:31 crc kubenswrapper[4578]: I1003 13:41:31.054082 4578 scope.go:117] "RemoveContainer" containerID="b9d656a3393eb5754f58714370a84898b144266c932eb399cd45b16ce7a45fa4" Oct 03 13:41:31 crc kubenswrapper[4578]: I1003 13:41:31.054701 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lsbxs"] Oct 03 13:41:31 crc kubenswrapper[4578]: I1003 13:41:31.064455 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lsbxs"] Oct 03 13:41:31 crc kubenswrapper[4578]: I1003 13:41:31.084154 4578 scope.go:117] "RemoveContainer" containerID="5f69bdeb0d4e61f9fea44b24213b6a747136bab8e0ac077ad55457b446dcc8ac" Oct 03 13:41:31 crc kubenswrapper[4578]: I1003 13:41:31.123466 4578 scope.go:117] "RemoveContainer" containerID="e2533197f40eea8fb13381b8c5d746cdef66e3db2066e291bcf6225eb969a6ea" Oct 03 13:41:31 crc kubenswrapper[4578]: E1003 13:41:31.123947 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2533197f40eea8fb13381b8c5d746cdef66e3db2066e291bcf6225eb969a6ea\": container with ID starting with e2533197f40eea8fb13381b8c5d746cdef66e3db2066e291bcf6225eb969a6ea not found: ID does not exist" containerID="e2533197f40eea8fb13381b8c5d746cdef66e3db2066e291bcf6225eb969a6ea" Oct 03 13:41:31 crc kubenswrapper[4578]: I1003 13:41:31.123982 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2533197f40eea8fb13381b8c5d746cdef66e3db2066e291bcf6225eb969a6ea"} err="failed to get container status \"e2533197f40eea8fb13381b8c5d746cdef66e3db2066e291bcf6225eb969a6ea\": rpc error: code = NotFound desc = could not find container \"e2533197f40eea8fb13381b8c5d746cdef66e3db2066e291bcf6225eb969a6ea\": container with ID starting with e2533197f40eea8fb13381b8c5d746cdef66e3db2066e291bcf6225eb969a6ea not found: ID does not exist" Oct 03 13:41:31 crc kubenswrapper[4578]: I1003 13:41:31.124004 4578 scope.go:117] "RemoveContainer" containerID="b9d656a3393eb5754f58714370a84898b144266c932eb399cd45b16ce7a45fa4" Oct 03 13:41:31 crc kubenswrapper[4578]: E1003 13:41:31.124252 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9d656a3393eb5754f58714370a84898b144266c932eb399cd45b16ce7a45fa4\": container with ID starting with b9d656a3393eb5754f58714370a84898b144266c932eb399cd45b16ce7a45fa4 not found: ID does not exist" containerID="b9d656a3393eb5754f58714370a84898b144266c932eb399cd45b16ce7a45fa4" Oct 03 13:41:31 crc kubenswrapper[4578]: I1003 13:41:31.124278 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9d656a3393eb5754f58714370a84898b144266c932eb399cd45b16ce7a45fa4"} err="failed to get container status \"b9d656a3393eb5754f58714370a84898b144266c932eb399cd45b16ce7a45fa4\": rpc error: code = NotFound desc = could not find container \"b9d656a3393eb5754f58714370a84898b144266c932eb399cd45b16ce7a45fa4\": container with ID starting with b9d656a3393eb5754f58714370a84898b144266c932eb399cd45b16ce7a45fa4 not found: ID does not exist" Oct 03 13:41:31 crc kubenswrapper[4578]: I1003 13:41:31.124291 4578 scope.go:117] "RemoveContainer" containerID="5f69bdeb0d4e61f9fea44b24213b6a747136bab8e0ac077ad55457b446dcc8ac" Oct 03 13:41:31 crc kubenswrapper[4578]: E1003 13:41:31.124477 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f69bdeb0d4e61f9fea44b24213b6a747136bab8e0ac077ad55457b446dcc8ac\": container with ID starting with 5f69bdeb0d4e61f9fea44b24213b6a747136bab8e0ac077ad55457b446dcc8ac not found: ID does not exist" containerID="5f69bdeb0d4e61f9fea44b24213b6a747136bab8e0ac077ad55457b446dcc8ac" Oct 03 13:41:31 crc kubenswrapper[4578]: I1003 13:41:31.124498 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f69bdeb0d4e61f9fea44b24213b6a747136bab8e0ac077ad55457b446dcc8ac"} err="failed to get container status \"5f69bdeb0d4e61f9fea44b24213b6a747136bab8e0ac077ad55457b446dcc8ac\": rpc error: code = NotFound desc = could not find container \"5f69bdeb0d4e61f9fea44b24213b6a747136bab8e0ac077ad55457b446dcc8ac\": container with ID starting with 5f69bdeb0d4e61f9fea44b24213b6a747136bab8e0ac077ad55457b446dcc8ac not found: ID does not exist" Oct 03 13:41:32 crc kubenswrapper[4578]: I1003 13:41:32.921095 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72c39a37-24c7-4e53-8f1b-b3a1794b9053" path="/var/lib/kubelet/pods/72c39a37-24c7-4e53-8f1b-b3a1794b9053/volumes" Oct 03 13:41:38 crc kubenswrapper[4578]: I1003 13:41:38.909695 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:41:38 crc kubenswrapper[4578]: E1003 13:41:38.910280 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:41:50 crc kubenswrapper[4578]: I1003 13:41:50.909040 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:41:50 crc kubenswrapper[4578]: E1003 13:41:50.909877 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:42:04 crc kubenswrapper[4578]: I1003 13:42:04.916520 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:42:04 crc kubenswrapper[4578]: E1003 13:42:04.917132 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:42:15 crc kubenswrapper[4578]: I1003 13:42:15.909142 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:42:15 crc kubenswrapper[4578]: E1003 13:42:15.909920 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:42:30 crc kubenswrapper[4578]: I1003 13:42:30.909402 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:42:30 crc kubenswrapper[4578]: E1003 13:42:30.910232 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:42:42 crc kubenswrapper[4578]: I1003 13:42:42.911063 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:42:42 crc kubenswrapper[4578]: E1003 13:42:42.912745 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.048217 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j"] Oct 03 13:42:45 crc kubenswrapper[4578]: E1003 13:42:45.050151 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72c39a37-24c7-4e53-8f1b-b3a1794b9053" containerName="extract-content" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.050177 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="72c39a37-24c7-4e53-8f1b-b3a1794b9053" containerName="extract-content" Oct 03 13:42:45 crc kubenswrapper[4578]: E1003 13:42:45.050204 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72c39a37-24c7-4e53-8f1b-b3a1794b9053" containerName="registry-server" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.050212 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="72c39a37-24c7-4e53-8f1b-b3a1794b9053" containerName="registry-server" Oct 03 13:42:45 crc kubenswrapper[4578]: E1003 13:42:45.050247 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72c39a37-24c7-4e53-8f1b-b3a1794b9053" containerName="extract-utilities" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.050256 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="72c39a37-24c7-4e53-8f1b-b3a1794b9053" containerName="extract-utilities" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.050477 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="72c39a37-24c7-4e53-8f1b-b3a1794b9053" containerName="registry-server" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.051307 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.055489 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-442mn" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.056009 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.056287 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.057099 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.062744 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.066689 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j"] Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.083823 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4df021ef-2d0b-4450-9c37-bd35443a3b1a-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k586j\" (UID: \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.084177 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs74d\" (UniqueName: \"kubernetes.io/projected/4df021ef-2d0b-4450-9c37-bd35443a3b1a-kube-api-access-zs74d\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k586j\" (UID: \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.084226 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4df021ef-2d0b-4450-9c37-bd35443a3b1a-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k586j\" (UID: \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.084276 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4df021ef-2d0b-4450-9c37-bd35443a3b1a-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k586j\" (UID: \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.084316 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4df021ef-2d0b-4450-9c37-bd35443a3b1a-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k586j\" (UID: \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.185825 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4df021ef-2d0b-4450-9c37-bd35443a3b1a-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k586j\" (UID: \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.185936 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs74d\" (UniqueName: \"kubernetes.io/projected/4df021ef-2d0b-4450-9c37-bd35443a3b1a-kube-api-access-zs74d\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k586j\" (UID: \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.185989 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4df021ef-2d0b-4450-9c37-bd35443a3b1a-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k586j\" (UID: \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.186084 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4df021ef-2d0b-4450-9c37-bd35443a3b1a-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k586j\" (UID: \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.186140 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4df021ef-2d0b-4450-9c37-bd35443a3b1a-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k586j\" (UID: \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.186846 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4df021ef-2d0b-4450-9c37-bd35443a3b1a-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k586j\" (UID: \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.196295 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4df021ef-2d0b-4450-9c37-bd35443a3b1a-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k586j\" (UID: \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.197351 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4df021ef-2d0b-4450-9c37-bd35443a3b1a-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k586j\" (UID: \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.197991 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4df021ef-2d0b-4450-9c37-bd35443a3b1a-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k586j\" (UID: \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.203228 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs74d\" (UniqueName: \"kubernetes.io/projected/4df021ef-2d0b-4450-9c37-bd35443a3b1a-kube-api-access-zs74d\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-k586j\" (UID: \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.396473 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.925705 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j"] Oct 03 13:42:45 crc kubenswrapper[4578]: I1003 13:42:45.933085 4578 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 13:42:46 crc kubenswrapper[4578]: I1003 13:42:46.646242 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" event={"ID":"4df021ef-2d0b-4450-9c37-bd35443a3b1a","Type":"ContainerStarted","Data":"fde0677c1770693c8a605835831159f910c40d58ba28828e315dd721fac155dd"} Oct 03 13:42:47 crc kubenswrapper[4578]: I1003 13:42:47.671517 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" event={"ID":"4df021ef-2d0b-4450-9c37-bd35443a3b1a","Type":"ContainerStarted","Data":"64e752768ce72e6be7e16257e90a0154d51ba1f8f55fb690889137abcc4509d5"} Oct 03 13:42:47 crc kubenswrapper[4578]: I1003 13:42:47.701610 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" podStartSLOduration=1.569661268 podStartE2EDuration="2.701593547s" podCreationTimestamp="2025-10-03 13:42:45 +0000 UTC" firstStartedPulling="2025-10-03 13:42:45.932835481 +0000 UTC m=+3101.731307675" lastFinishedPulling="2025-10-03 13:42:47.06476776 +0000 UTC m=+3102.863239954" observedRunningTime="2025-10-03 13:42:47.700182523 +0000 UTC m=+3103.498654717" watchObservedRunningTime="2025-10-03 13:42:47.701593547 +0000 UTC m=+3103.500065731" Oct 03 13:42:53 crc kubenswrapper[4578]: I1003 13:42:53.909320 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:42:53 crc kubenswrapper[4578]: E1003 13:42:53.910149 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:43:08 crc kubenswrapper[4578]: I1003 13:43:08.909031 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:43:08 crc kubenswrapper[4578]: E1003 13:43:08.909802 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:43:18 crc kubenswrapper[4578]: I1003 13:43:18.924406 4578 generic.go:334] "Generic (PLEG): container finished" podID="4df021ef-2d0b-4450-9c37-bd35443a3b1a" containerID="64e752768ce72e6be7e16257e90a0154d51ba1f8f55fb690889137abcc4509d5" exitCode=2 Oct 03 13:43:18 crc kubenswrapper[4578]: I1003 13:43:18.925724 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" event={"ID":"4df021ef-2d0b-4450-9c37-bd35443a3b1a","Type":"ContainerDied","Data":"64e752768ce72e6be7e16257e90a0154d51ba1f8f55fb690889137abcc4509d5"} Oct 03 13:43:20 crc kubenswrapper[4578]: I1003 13:43:20.311574 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" Oct 03 13:43:20 crc kubenswrapper[4578]: I1003 13:43:20.437446 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zs74d\" (UniqueName: \"kubernetes.io/projected/4df021ef-2d0b-4450-9c37-bd35443a3b1a-kube-api-access-zs74d\") pod \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\" (UID: \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\") " Oct 03 13:43:20 crc kubenswrapper[4578]: I1003 13:43:20.437548 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4df021ef-2d0b-4450-9c37-bd35443a3b1a-inventory\") pod \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\" (UID: \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\") " Oct 03 13:43:20 crc kubenswrapper[4578]: I1003 13:43:20.437621 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4df021ef-2d0b-4450-9c37-bd35443a3b1a-ssh-key\") pod \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\" (UID: \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\") " Oct 03 13:43:20 crc kubenswrapper[4578]: I1003 13:43:20.437749 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4df021ef-2d0b-4450-9c37-bd35443a3b1a-ovn-combined-ca-bundle\") pod \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\" (UID: \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\") " Oct 03 13:43:20 crc kubenswrapper[4578]: I1003 13:43:20.437780 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4df021ef-2d0b-4450-9c37-bd35443a3b1a-ovncontroller-config-0\") pod \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\" (UID: \"4df021ef-2d0b-4450-9c37-bd35443a3b1a\") " Oct 03 13:43:20 crc kubenswrapper[4578]: I1003 13:43:20.443979 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4df021ef-2d0b-4450-9c37-bd35443a3b1a-kube-api-access-zs74d" (OuterVolumeSpecName: "kube-api-access-zs74d") pod "4df021ef-2d0b-4450-9c37-bd35443a3b1a" (UID: "4df021ef-2d0b-4450-9c37-bd35443a3b1a"). InnerVolumeSpecName "kube-api-access-zs74d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:43:20 crc kubenswrapper[4578]: I1003 13:43:20.445947 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4df021ef-2d0b-4450-9c37-bd35443a3b1a-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "4df021ef-2d0b-4450-9c37-bd35443a3b1a" (UID: "4df021ef-2d0b-4450-9c37-bd35443a3b1a"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:43:20 crc kubenswrapper[4578]: I1003 13:43:20.464797 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4df021ef-2d0b-4450-9c37-bd35443a3b1a-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "4df021ef-2d0b-4450-9c37-bd35443a3b1a" (UID: "4df021ef-2d0b-4450-9c37-bd35443a3b1a"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:43:20 crc kubenswrapper[4578]: I1003 13:43:20.468708 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4df021ef-2d0b-4450-9c37-bd35443a3b1a-inventory" (OuterVolumeSpecName: "inventory") pod "4df021ef-2d0b-4450-9c37-bd35443a3b1a" (UID: "4df021ef-2d0b-4450-9c37-bd35443a3b1a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:43:20 crc kubenswrapper[4578]: I1003 13:43:20.472096 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4df021ef-2d0b-4450-9c37-bd35443a3b1a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4df021ef-2d0b-4450-9c37-bd35443a3b1a" (UID: "4df021ef-2d0b-4450-9c37-bd35443a3b1a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:43:20 crc kubenswrapper[4578]: I1003 13:43:20.539727 4578 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4df021ef-2d0b-4450-9c37-bd35443a3b1a-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 13:43:20 crc kubenswrapper[4578]: I1003 13:43:20.539767 4578 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4df021ef-2d0b-4450-9c37-bd35443a3b1a-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 03 13:43:20 crc kubenswrapper[4578]: I1003 13:43:20.539777 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zs74d\" (UniqueName: \"kubernetes.io/projected/4df021ef-2d0b-4450-9c37-bd35443a3b1a-kube-api-access-zs74d\") on node \"crc\" DevicePath \"\"" Oct 03 13:43:20 crc kubenswrapper[4578]: I1003 13:43:20.539787 4578 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4df021ef-2d0b-4450-9c37-bd35443a3b1a-inventory\") on node \"crc\" DevicePath \"\"" Oct 03 13:43:20 crc kubenswrapper[4578]: I1003 13:43:20.539796 4578 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4df021ef-2d0b-4450-9c37-bd35443a3b1a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 03 13:43:20 crc kubenswrapper[4578]: I1003 13:43:20.948353 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" event={"ID":"4df021ef-2d0b-4450-9c37-bd35443a3b1a","Type":"ContainerDied","Data":"fde0677c1770693c8a605835831159f910c40d58ba28828e315dd721fac155dd"} Oct 03 13:43:20 crc kubenswrapper[4578]: I1003 13:43:20.948399 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fde0677c1770693c8a605835831159f910c40d58ba28828e315dd721fac155dd" Oct 03 13:43:20 crc kubenswrapper[4578]: I1003 13:43:20.948402 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-k586j" Oct 03 13:43:23 crc kubenswrapper[4578]: I1003 13:43:23.910159 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:43:23 crc kubenswrapper[4578]: E1003 13:43:23.910866 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:43:35 crc kubenswrapper[4578]: I1003 13:43:35.909882 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:43:35 crc kubenswrapper[4578]: E1003 13:43:35.910892 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:43:50 crc kubenswrapper[4578]: I1003 13:43:50.912807 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:43:50 crc kubenswrapper[4578]: E1003 13:43:50.913686 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:44:03 crc kubenswrapper[4578]: I1003 13:44:03.908697 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:44:03 crc kubenswrapper[4578]: E1003 13:44:03.909446 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:44:15 crc kubenswrapper[4578]: I1003 13:44:15.910148 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:44:16 crc kubenswrapper[4578]: I1003 13:44:16.414606 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"63284c2a61274b6348f80287108612ab5c7b2a261fcd4d7590b033486c46447c"} Oct 03 13:45:00 crc kubenswrapper[4578]: I1003 13:45:00.159012 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324985-6th97"] Oct 03 13:45:00 crc kubenswrapper[4578]: E1003 13:45:00.160045 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4df021ef-2d0b-4450-9c37-bd35443a3b1a" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:45:00 crc kubenswrapper[4578]: I1003 13:45:00.160061 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="4df021ef-2d0b-4450-9c37-bd35443a3b1a" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:45:00 crc kubenswrapper[4578]: I1003 13:45:00.160299 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="4df021ef-2d0b-4450-9c37-bd35443a3b1a" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 03 13:45:00 crc kubenswrapper[4578]: I1003 13:45:00.161063 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-6th97" Oct 03 13:45:00 crc kubenswrapper[4578]: I1003 13:45:00.167293 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 13:45:00 crc kubenswrapper[4578]: I1003 13:45:00.169844 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 13:45:00 crc kubenswrapper[4578]: I1003 13:45:00.180459 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324985-6th97"] Oct 03 13:45:00 crc kubenswrapper[4578]: I1003 13:45:00.324262 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a19bc9a1-839c-470a-ab05-5079ed70f8c8-secret-volume\") pod \"collect-profiles-29324985-6th97\" (UID: \"a19bc9a1-839c-470a-ab05-5079ed70f8c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-6th97" Oct 03 13:45:00 crc kubenswrapper[4578]: I1003 13:45:00.324349 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a19bc9a1-839c-470a-ab05-5079ed70f8c8-config-volume\") pod \"collect-profiles-29324985-6th97\" (UID: \"a19bc9a1-839c-470a-ab05-5079ed70f8c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-6th97" Oct 03 13:45:00 crc kubenswrapper[4578]: I1003 13:45:00.324391 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nwnx\" (UniqueName: \"kubernetes.io/projected/a19bc9a1-839c-470a-ab05-5079ed70f8c8-kube-api-access-5nwnx\") pod \"collect-profiles-29324985-6th97\" (UID: \"a19bc9a1-839c-470a-ab05-5079ed70f8c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-6th97" Oct 03 13:45:00 crc kubenswrapper[4578]: I1003 13:45:00.427516 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nwnx\" (UniqueName: \"kubernetes.io/projected/a19bc9a1-839c-470a-ab05-5079ed70f8c8-kube-api-access-5nwnx\") pod \"collect-profiles-29324985-6th97\" (UID: \"a19bc9a1-839c-470a-ab05-5079ed70f8c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-6th97" Oct 03 13:45:00 crc kubenswrapper[4578]: I1003 13:45:00.427741 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a19bc9a1-839c-470a-ab05-5079ed70f8c8-secret-volume\") pod \"collect-profiles-29324985-6th97\" (UID: \"a19bc9a1-839c-470a-ab05-5079ed70f8c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-6th97" Oct 03 13:45:00 crc kubenswrapper[4578]: I1003 13:45:00.427804 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a19bc9a1-839c-470a-ab05-5079ed70f8c8-config-volume\") pod \"collect-profiles-29324985-6th97\" (UID: \"a19bc9a1-839c-470a-ab05-5079ed70f8c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-6th97" Oct 03 13:45:00 crc kubenswrapper[4578]: I1003 13:45:00.428650 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a19bc9a1-839c-470a-ab05-5079ed70f8c8-config-volume\") pod \"collect-profiles-29324985-6th97\" (UID: \"a19bc9a1-839c-470a-ab05-5079ed70f8c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-6th97" Oct 03 13:45:00 crc kubenswrapper[4578]: I1003 13:45:00.433899 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a19bc9a1-839c-470a-ab05-5079ed70f8c8-secret-volume\") pod \"collect-profiles-29324985-6th97\" (UID: \"a19bc9a1-839c-470a-ab05-5079ed70f8c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-6th97" Oct 03 13:45:00 crc kubenswrapper[4578]: I1003 13:45:00.448992 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nwnx\" (UniqueName: \"kubernetes.io/projected/a19bc9a1-839c-470a-ab05-5079ed70f8c8-kube-api-access-5nwnx\") pod \"collect-profiles-29324985-6th97\" (UID: \"a19bc9a1-839c-470a-ab05-5079ed70f8c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-6th97" Oct 03 13:45:00 crc kubenswrapper[4578]: I1003 13:45:00.491540 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-6th97" Oct 03 13:45:00 crc kubenswrapper[4578]: I1003 13:45:00.943199 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324985-6th97"] Oct 03 13:45:01 crc kubenswrapper[4578]: I1003 13:45:01.832135 4578 generic.go:334] "Generic (PLEG): container finished" podID="a19bc9a1-839c-470a-ab05-5079ed70f8c8" containerID="5d927925f57ad0b21d838ceac4208b0edd0ed9d69625eafb06500796481af091" exitCode=0 Oct 03 13:45:01 crc kubenswrapper[4578]: I1003 13:45:01.832262 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-6th97" event={"ID":"a19bc9a1-839c-470a-ab05-5079ed70f8c8","Type":"ContainerDied","Data":"5d927925f57ad0b21d838ceac4208b0edd0ed9d69625eafb06500796481af091"} Oct 03 13:45:01 crc kubenswrapper[4578]: I1003 13:45:01.832475 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-6th97" event={"ID":"a19bc9a1-839c-470a-ab05-5079ed70f8c8","Type":"ContainerStarted","Data":"ebda4abb8bea23f382d05398c4009fd6e7720bd31667529a4526bc24be8d156e"} Oct 03 13:45:03 crc kubenswrapper[4578]: I1003 13:45:03.147917 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-6th97" Oct 03 13:45:03 crc kubenswrapper[4578]: I1003 13:45:03.182090 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nwnx\" (UniqueName: \"kubernetes.io/projected/a19bc9a1-839c-470a-ab05-5079ed70f8c8-kube-api-access-5nwnx\") pod \"a19bc9a1-839c-470a-ab05-5079ed70f8c8\" (UID: \"a19bc9a1-839c-470a-ab05-5079ed70f8c8\") " Oct 03 13:45:03 crc kubenswrapper[4578]: I1003 13:45:03.182231 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a19bc9a1-839c-470a-ab05-5079ed70f8c8-config-volume\") pod \"a19bc9a1-839c-470a-ab05-5079ed70f8c8\" (UID: \"a19bc9a1-839c-470a-ab05-5079ed70f8c8\") " Oct 03 13:45:03 crc kubenswrapper[4578]: I1003 13:45:03.182281 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a19bc9a1-839c-470a-ab05-5079ed70f8c8-secret-volume\") pod \"a19bc9a1-839c-470a-ab05-5079ed70f8c8\" (UID: \"a19bc9a1-839c-470a-ab05-5079ed70f8c8\") " Oct 03 13:45:03 crc kubenswrapper[4578]: I1003 13:45:03.182867 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a19bc9a1-839c-470a-ab05-5079ed70f8c8-config-volume" (OuterVolumeSpecName: "config-volume") pod "a19bc9a1-839c-470a-ab05-5079ed70f8c8" (UID: "a19bc9a1-839c-470a-ab05-5079ed70f8c8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 13:45:03 crc kubenswrapper[4578]: I1003 13:45:03.190117 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a19bc9a1-839c-470a-ab05-5079ed70f8c8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a19bc9a1-839c-470a-ab05-5079ed70f8c8" (UID: "a19bc9a1-839c-470a-ab05-5079ed70f8c8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 13:45:03 crc kubenswrapper[4578]: I1003 13:45:03.190417 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a19bc9a1-839c-470a-ab05-5079ed70f8c8-kube-api-access-5nwnx" (OuterVolumeSpecName: "kube-api-access-5nwnx") pod "a19bc9a1-839c-470a-ab05-5079ed70f8c8" (UID: "a19bc9a1-839c-470a-ab05-5079ed70f8c8"). InnerVolumeSpecName "kube-api-access-5nwnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:45:03 crc kubenswrapper[4578]: I1003 13:45:03.284335 4578 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a19bc9a1-839c-470a-ab05-5079ed70f8c8-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 13:45:03 crc kubenswrapper[4578]: I1003 13:45:03.284374 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nwnx\" (UniqueName: \"kubernetes.io/projected/a19bc9a1-839c-470a-ab05-5079ed70f8c8-kube-api-access-5nwnx\") on node \"crc\" DevicePath \"\"" Oct 03 13:45:03 crc kubenswrapper[4578]: I1003 13:45:03.284388 4578 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a19bc9a1-839c-470a-ab05-5079ed70f8c8-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 13:45:03 crc kubenswrapper[4578]: I1003 13:45:03.859011 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-6th97" event={"ID":"a19bc9a1-839c-470a-ab05-5079ed70f8c8","Type":"ContainerDied","Data":"ebda4abb8bea23f382d05398c4009fd6e7720bd31667529a4526bc24be8d156e"} Oct 03 13:45:03 crc kubenswrapper[4578]: I1003 13:45:03.859058 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebda4abb8bea23f382d05398c4009fd6e7720bd31667529a4526bc24be8d156e" Oct 03 13:45:03 crc kubenswrapper[4578]: I1003 13:45:03.859123 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29324985-6th97" Oct 03 13:45:04 crc kubenswrapper[4578]: I1003 13:45:04.226363 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z"] Oct 03 13:45:04 crc kubenswrapper[4578]: I1003 13:45:04.234695 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324940-gpv7z"] Oct 03 13:45:04 crc kubenswrapper[4578]: I1003 13:45:04.928582 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d777f4d5-fab6-4033-9200-e7cce416dab9" path="/var/lib/kubelet/pods/d777f4d5-fab6-4033-9200-e7cce416dab9/volumes" Oct 03 13:45:09 crc kubenswrapper[4578]: I1003 13:45:09.845862 4578 scope.go:117] "RemoveContainer" containerID="7587bf723aa48d53704b41dd167508e205437c8b6ed9ac61bf3b4b21fdd01e82" Oct 03 13:46:35 crc kubenswrapper[4578]: I1003 13:46:35.091138 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:46:35 crc kubenswrapper[4578]: I1003 13:46:35.091808 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:47:05 crc kubenswrapper[4578]: I1003 13:47:05.091901 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:47:05 crc kubenswrapper[4578]: I1003 13:47:05.092827 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:47:35 crc kubenswrapper[4578]: I1003 13:47:35.091196 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:47:35 crc kubenswrapper[4578]: I1003 13:47:35.091754 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:47:35 crc kubenswrapper[4578]: I1003 13:47:35.091808 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 13:47:35 crc kubenswrapper[4578]: I1003 13:47:35.092494 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"63284c2a61274b6348f80287108612ab5c7b2a261fcd4d7590b033486c46447c"} pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:47:35 crc kubenswrapper[4578]: I1003 13:47:35.092543 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" containerID="cri-o://63284c2a61274b6348f80287108612ab5c7b2a261fcd4d7590b033486c46447c" gracePeriod=600 Oct 03 13:47:35 crc kubenswrapper[4578]: I1003 13:47:35.369537 4578 generic.go:334] "Generic (PLEG): container finished" podID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerID="63284c2a61274b6348f80287108612ab5c7b2a261fcd4d7590b033486c46447c" exitCode=0 Oct 03 13:47:35 crc kubenswrapper[4578]: I1003 13:47:35.369603 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerDied","Data":"63284c2a61274b6348f80287108612ab5c7b2a261fcd4d7590b033486c46447c"} Oct 03 13:47:35 crc kubenswrapper[4578]: I1003 13:47:35.369894 4578 scope.go:117] "RemoveContainer" containerID="f9c66ced7b9b39a9d28f9e3669a1556638cd50cf6fc749e111be29ca6e1bf406" Oct 03 13:47:36 crc kubenswrapper[4578]: I1003 13:47:36.379057 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e"} Oct 03 13:48:07 crc kubenswrapper[4578]: I1003 13:48:07.551252 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8jx6r"] Oct 03 13:48:07 crc kubenswrapper[4578]: E1003 13:48:07.552247 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a19bc9a1-839c-470a-ab05-5079ed70f8c8" containerName="collect-profiles" Oct 03 13:48:07 crc kubenswrapper[4578]: I1003 13:48:07.552258 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a19bc9a1-839c-470a-ab05-5079ed70f8c8" containerName="collect-profiles" Oct 03 13:48:07 crc kubenswrapper[4578]: I1003 13:48:07.552486 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="a19bc9a1-839c-470a-ab05-5079ed70f8c8" containerName="collect-profiles" Oct 03 13:48:07 crc kubenswrapper[4578]: I1003 13:48:07.553745 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8jx6r" Oct 03 13:48:07 crc kubenswrapper[4578]: I1003 13:48:07.565132 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8jx6r"] Oct 03 13:48:07 crc kubenswrapper[4578]: I1003 13:48:07.642475 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/245a6647-a604-425b-8c8d-865f3bfc9fb6-catalog-content\") pod \"certified-operators-8jx6r\" (UID: \"245a6647-a604-425b-8c8d-865f3bfc9fb6\") " pod="openshift-marketplace/certified-operators-8jx6r" Oct 03 13:48:07 crc kubenswrapper[4578]: I1003 13:48:07.642571 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/245a6647-a604-425b-8c8d-865f3bfc9fb6-utilities\") pod \"certified-operators-8jx6r\" (UID: \"245a6647-a604-425b-8c8d-865f3bfc9fb6\") " pod="openshift-marketplace/certified-operators-8jx6r" Oct 03 13:48:07 crc kubenswrapper[4578]: I1003 13:48:07.642664 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt2rn\" (UniqueName: \"kubernetes.io/projected/245a6647-a604-425b-8c8d-865f3bfc9fb6-kube-api-access-bt2rn\") pod \"certified-operators-8jx6r\" (UID: \"245a6647-a604-425b-8c8d-865f3bfc9fb6\") " pod="openshift-marketplace/certified-operators-8jx6r" Oct 03 13:48:07 crc kubenswrapper[4578]: I1003 13:48:07.744698 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/245a6647-a604-425b-8c8d-865f3bfc9fb6-catalog-content\") pod \"certified-operators-8jx6r\" (UID: \"245a6647-a604-425b-8c8d-865f3bfc9fb6\") " pod="openshift-marketplace/certified-operators-8jx6r" Oct 03 13:48:07 crc kubenswrapper[4578]: I1003 13:48:07.744763 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/245a6647-a604-425b-8c8d-865f3bfc9fb6-utilities\") pod \"certified-operators-8jx6r\" (UID: \"245a6647-a604-425b-8c8d-865f3bfc9fb6\") " pod="openshift-marketplace/certified-operators-8jx6r" Oct 03 13:48:07 crc kubenswrapper[4578]: I1003 13:48:07.744795 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt2rn\" (UniqueName: \"kubernetes.io/projected/245a6647-a604-425b-8c8d-865f3bfc9fb6-kube-api-access-bt2rn\") pod \"certified-operators-8jx6r\" (UID: \"245a6647-a604-425b-8c8d-865f3bfc9fb6\") " pod="openshift-marketplace/certified-operators-8jx6r" Oct 03 13:48:07 crc kubenswrapper[4578]: I1003 13:48:07.745355 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/245a6647-a604-425b-8c8d-865f3bfc9fb6-utilities\") pod \"certified-operators-8jx6r\" (UID: \"245a6647-a604-425b-8c8d-865f3bfc9fb6\") " pod="openshift-marketplace/certified-operators-8jx6r" Oct 03 13:48:07 crc kubenswrapper[4578]: I1003 13:48:07.745368 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/245a6647-a604-425b-8c8d-865f3bfc9fb6-catalog-content\") pod \"certified-operators-8jx6r\" (UID: \"245a6647-a604-425b-8c8d-865f3bfc9fb6\") " pod="openshift-marketplace/certified-operators-8jx6r" Oct 03 13:48:07 crc kubenswrapper[4578]: I1003 13:48:07.763050 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt2rn\" (UniqueName: \"kubernetes.io/projected/245a6647-a604-425b-8c8d-865f3bfc9fb6-kube-api-access-bt2rn\") pod \"certified-operators-8jx6r\" (UID: \"245a6647-a604-425b-8c8d-865f3bfc9fb6\") " pod="openshift-marketplace/certified-operators-8jx6r" Oct 03 13:48:07 crc kubenswrapper[4578]: I1003 13:48:07.877181 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8jx6r" Oct 03 13:48:08 crc kubenswrapper[4578]: I1003 13:48:08.418537 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8jx6r"] Oct 03 13:48:08 crc kubenswrapper[4578]: I1003 13:48:08.667620 4578 generic.go:334] "Generic (PLEG): container finished" podID="245a6647-a604-425b-8c8d-865f3bfc9fb6" containerID="fb29322976620a4c039fd60a0c740b192eaefa75c8c4f3712d7822dbc41d97d2" exitCode=0 Oct 03 13:48:08 crc kubenswrapper[4578]: I1003 13:48:08.667681 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8jx6r" event={"ID":"245a6647-a604-425b-8c8d-865f3bfc9fb6","Type":"ContainerDied","Data":"fb29322976620a4c039fd60a0c740b192eaefa75c8c4f3712d7822dbc41d97d2"} Oct 03 13:48:08 crc kubenswrapper[4578]: I1003 13:48:08.667708 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8jx6r" event={"ID":"245a6647-a604-425b-8c8d-865f3bfc9fb6","Type":"ContainerStarted","Data":"ccbee4d22fd65e3d026d53343f5b8e12e547e9b67f9675a92f6fc9ed87c230e9"} Oct 03 13:48:09 crc kubenswrapper[4578]: I1003 13:48:09.680594 4578 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 13:48:11 crc kubenswrapper[4578]: I1003 13:48:11.703199 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8jx6r" event={"ID":"245a6647-a604-425b-8c8d-865f3bfc9fb6","Type":"ContainerStarted","Data":"869c46846cee54e9ad40480ad99becbf8871caa934e48d9ff3a1b4d7d3ae1f6b"} Oct 03 13:48:14 crc kubenswrapper[4578]: I1003 13:48:14.728530 4578 generic.go:334] "Generic (PLEG): container finished" podID="245a6647-a604-425b-8c8d-865f3bfc9fb6" containerID="869c46846cee54e9ad40480ad99becbf8871caa934e48d9ff3a1b4d7d3ae1f6b" exitCode=0 Oct 03 13:48:14 crc kubenswrapper[4578]: I1003 13:48:14.728570 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8jx6r" event={"ID":"245a6647-a604-425b-8c8d-865f3bfc9fb6","Type":"ContainerDied","Data":"869c46846cee54e9ad40480ad99becbf8871caa934e48d9ff3a1b4d7d3ae1f6b"} Oct 03 13:48:15 crc kubenswrapper[4578]: I1003 13:48:15.739431 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8jx6r" event={"ID":"245a6647-a604-425b-8c8d-865f3bfc9fb6","Type":"ContainerStarted","Data":"6e019e5bed99bf9bff7dd92745bd1c23c57b274dd7e840ae85d69064b09c77ea"} Oct 03 13:48:15 crc kubenswrapper[4578]: I1003 13:48:15.764969 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8jx6r" podStartSLOduration=2.937881028 podStartE2EDuration="8.764954771s" podCreationTimestamp="2025-10-03 13:48:07 +0000 UTC" firstStartedPulling="2025-10-03 13:48:09.680356875 +0000 UTC m=+3425.478829049" lastFinishedPulling="2025-10-03 13:48:15.507430608 +0000 UTC m=+3431.305902792" observedRunningTime="2025-10-03 13:48:15.761147652 +0000 UTC m=+3431.559619836" watchObservedRunningTime="2025-10-03 13:48:15.764954771 +0000 UTC m=+3431.563426955" Oct 03 13:48:17 crc kubenswrapper[4578]: I1003 13:48:17.879060 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8jx6r" Oct 03 13:48:17 crc kubenswrapper[4578]: I1003 13:48:17.880410 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8jx6r" Oct 03 13:48:17 crc kubenswrapper[4578]: I1003 13:48:17.929376 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8jx6r" Oct 03 13:48:27 crc kubenswrapper[4578]: I1003 13:48:27.922509 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8jx6r" Oct 03 13:48:27 crc kubenswrapper[4578]: I1003 13:48:27.974004 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8jx6r"] Oct 03 13:48:28 crc kubenswrapper[4578]: I1003 13:48:28.853258 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8jx6r" podUID="245a6647-a604-425b-8c8d-865f3bfc9fb6" containerName="registry-server" containerID="cri-o://6e019e5bed99bf9bff7dd92745bd1c23c57b274dd7e840ae85d69064b09c77ea" gracePeriod=2 Oct 03 13:48:29 crc kubenswrapper[4578]: I1003 13:48:29.289168 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8jx6r" Oct 03 13:48:29 crc kubenswrapper[4578]: I1003 13:48:29.449013 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/245a6647-a604-425b-8c8d-865f3bfc9fb6-utilities\") pod \"245a6647-a604-425b-8c8d-865f3bfc9fb6\" (UID: \"245a6647-a604-425b-8c8d-865f3bfc9fb6\") " Oct 03 13:48:29 crc kubenswrapper[4578]: I1003 13:48:29.449405 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bt2rn\" (UniqueName: \"kubernetes.io/projected/245a6647-a604-425b-8c8d-865f3bfc9fb6-kube-api-access-bt2rn\") pod \"245a6647-a604-425b-8c8d-865f3bfc9fb6\" (UID: \"245a6647-a604-425b-8c8d-865f3bfc9fb6\") " Oct 03 13:48:29 crc kubenswrapper[4578]: I1003 13:48:29.449478 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/245a6647-a604-425b-8c8d-865f3bfc9fb6-catalog-content\") pod \"245a6647-a604-425b-8c8d-865f3bfc9fb6\" (UID: \"245a6647-a604-425b-8c8d-865f3bfc9fb6\") " Oct 03 13:48:29 crc kubenswrapper[4578]: I1003 13:48:29.450303 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/245a6647-a604-425b-8c8d-865f3bfc9fb6-utilities" (OuterVolumeSpecName: "utilities") pod "245a6647-a604-425b-8c8d-865f3bfc9fb6" (UID: "245a6647-a604-425b-8c8d-865f3bfc9fb6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:48:29 crc kubenswrapper[4578]: I1003 13:48:29.450420 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/245a6647-a604-425b-8c8d-865f3bfc9fb6-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:48:29 crc kubenswrapper[4578]: I1003 13:48:29.456080 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/245a6647-a604-425b-8c8d-865f3bfc9fb6-kube-api-access-bt2rn" (OuterVolumeSpecName: "kube-api-access-bt2rn") pod "245a6647-a604-425b-8c8d-865f3bfc9fb6" (UID: "245a6647-a604-425b-8c8d-865f3bfc9fb6"). InnerVolumeSpecName "kube-api-access-bt2rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:48:29 crc kubenswrapper[4578]: I1003 13:48:29.489520 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/245a6647-a604-425b-8c8d-865f3bfc9fb6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "245a6647-a604-425b-8c8d-865f3bfc9fb6" (UID: "245a6647-a604-425b-8c8d-865f3bfc9fb6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:48:29 crc kubenswrapper[4578]: I1003 13:48:29.552091 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bt2rn\" (UniqueName: \"kubernetes.io/projected/245a6647-a604-425b-8c8d-865f3bfc9fb6-kube-api-access-bt2rn\") on node \"crc\" DevicePath \"\"" Oct 03 13:48:29 crc kubenswrapper[4578]: I1003 13:48:29.552120 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/245a6647-a604-425b-8c8d-865f3bfc9fb6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:48:29 crc kubenswrapper[4578]: I1003 13:48:29.879899 4578 generic.go:334] "Generic (PLEG): container finished" podID="245a6647-a604-425b-8c8d-865f3bfc9fb6" containerID="6e019e5bed99bf9bff7dd92745bd1c23c57b274dd7e840ae85d69064b09c77ea" exitCode=0 Oct 03 13:48:29 crc kubenswrapper[4578]: I1003 13:48:29.879946 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8jx6r" event={"ID":"245a6647-a604-425b-8c8d-865f3bfc9fb6","Type":"ContainerDied","Data":"6e019e5bed99bf9bff7dd92745bd1c23c57b274dd7e840ae85d69064b09c77ea"} Oct 03 13:48:29 crc kubenswrapper[4578]: I1003 13:48:29.879973 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8jx6r" event={"ID":"245a6647-a604-425b-8c8d-865f3bfc9fb6","Type":"ContainerDied","Data":"ccbee4d22fd65e3d026d53343f5b8e12e547e9b67f9675a92f6fc9ed87c230e9"} Oct 03 13:48:29 crc kubenswrapper[4578]: I1003 13:48:29.879991 4578 scope.go:117] "RemoveContainer" containerID="6e019e5bed99bf9bff7dd92745bd1c23c57b274dd7e840ae85d69064b09c77ea" Oct 03 13:48:29 crc kubenswrapper[4578]: I1003 13:48:29.880160 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8jx6r" Oct 03 13:48:29 crc kubenswrapper[4578]: I1003 13:48:29.967692 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8jx6r"] Oct 03 13:48:29 crc kubenswrapper[4578]: I1003 13:48:29.974867 4578 scope.go:117] "RemoveContainer" containerID="869c46846cee54e9ad40480ad99becbf8871caa934e48d9ff3a1b4d7d3ae1f6b" Oct 03 13:48:29 crc kubenswrapper[4578]: I1003 13:48:29.983447 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8jx6r"] Oct 03 13:48:30 crc kubenswrapper[4578]: I1003 13:48:30.042956 4578 scope.go:117] "RemoveContainer" containerID="fb29322976620a4c039fd60a0c740b192eaefa75c8c4f3712d7822dbc41d97d2" Oct 03 13:48:30 crc kubenswrapper[4578]: I1003 13:48:30.095813 4578 scope.go:117] "RemoveContainer" containerID="6e019e5bed99bf9bff7dd92745bd1c23c57b274dd7e840ae85d69064b09c77ea" Oct 03 13:48:30 crc kubenswrapper[4578]: E1003 13:48:30.098778 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e019e5bed99bf9bff7dd92745bd1c23c57b274dd7e840ae85d69064b09c77ea\": container with ID starting with 6e019e5bed99bf9bff7dd92745bd1c23c57b274dd7e840ae85d69064b09c77ea not found: ID does not exist" containerID="6e019e5bed99bf9bff7dd92745bd1c23c57b274dd7e840ae85d69064b09c77ea" Oct 03 13:48:30 crc kubenswrapper[4578]: I1003 13:48:30.098822 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e019e5bed99bf9bff7dd92745bd1c23c57b274dd7e840ae85d69064b09c77ea"} err="failed to get container status \"6e019e5bed99bf9bff7dd92745bd1c23c57b274dd7e840ae85d69064b09c77ea\": rpc error: code = NotFound desc = could not find container \"6e019e5bed99bf9bff7dd92745bd1c23c57b274dd7e840ae85d69064b09c77ea\": container with ID starting with 6e019e5bed99bf9bff7dd92745bd1c23c57b274dd7e840ae85d69064b09c77ea not found: ID does not exist" Oct 03 13:48:30 crc kubenswrapper[4578]: I1003 13:48:30.098846 4578 scope.go:117] "RemoveContainer" containerID="869c46846cee54e9ad40480ad99becbf8871caa934e48d9ff3a1b4d7d3ae1f6b" Oct 03 13:48:30 crc kubenswrapper[4578]: E1003 13:48:30.099223 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"869c46846cee54e9ad40480ad99becbf8871caa934e48d9ff3a1b4d7d3ae1f6b\": container with ID starting with 869c46846cee54e9ad40480ad99becbf8871caa934e48d9ff3a1b4d7d3ae1f6b not found: ID does not exist" containerID="869c46846cee54e9ad40480ad99becbf8871caa934e48d9ff3a1b4d7d3ae1f6b" Oct 03 13:48:30 crc kubenswrapper[4578]: I1003 13:48:30.099242 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"869c46846cee54e9ad40480ad99becbf8871caa934e48d9ff3a1b4d7d3ae1f6b"} err="failed to get container status \"869c46846cee54e9ad40480ad99becbf8871caa934e48d9ff3a1b4d7d3ae1f6b\": rpc error: code = NotFound desc = could not find container \"869c46846cee54e9ad40480ad99becbf8871caa934e48d9ff3a1b4d7d3ae1f6b\": container with ID starting with 869c46846cee54e9ad40480ad99becbf8871caa934e48d9ff3a1b4d7d3ae1f6b not found: ID does not exist" Oct 03 13:48:30 crc kubenswrapper[4578]: I1003 13:48:30.099256 4578 scope.go:117] "RemoveContainer" containerID="fb29322976620a4c039fd60a0c740b192eaefa75c8c4f3712d7822dbc41d97d2" Oct 03 13:48:30 crc kubenswrapper[4578]: E1003 13:48:30.099566 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb29322976620a4c039fd60a0c740b192eaefa75c8c4f3712d7822dbc41d97d2\": container with ID starting with fb29322976620a4c039fd60a0c740b192eaefa75c8c4f3712d7822dbc41d97d2 not found: ID does not exist" containerID="fb29322976620a4c039fd60a0c740b192eaefa75c8c4f3712d7822dbc41d97d2" Oct 03 13:48:30 crc kubenswrapper[4578]: I1003 13:48:30.099591 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb29322976620a4c039fd60a0c740b192eaefa75c8c4f3712d7822dbc41d97d2"} err="failed to get container status \"fb29322976620a4c039fd60a0c740b192eaefa75c8c4f3712d7822dbc41d97d2\": rpc error: code = NotFound desc = could not find container \"fb29322976620a4c039fd60a0c740b192eaefa75c8c4f3712d7822dbc41d97d2\": container with ID starting with fb29322976620a4c039fd60a0c740b192eaefa75c8c4f3712d7822dbc41d97d2 not found: ID does not exist" Oct 03 13:48:30 crc kubenswrapper[4578]: I1003 13:48:30.918833 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="245a6647-a604-425b-8c8d-865f3bfc9fb6" path="/var/lib/kubelet/pods/245a6647-a604-425b-8c8d-865f3bfc9fb6/volumes" Oct 03 13:49:35 crc kubenswrapper[4578]: I1003 13:49:35.091992 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:49:35 crc kubenswrapper[4578]: I1003 13:49:35.092805 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:50:05 crc kubenswrapper[4578]: I1003 13:50:05.091925 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:50:05 crc kubenswrapper[4578]: I1003 13:50:05.092468 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:50:09 crc kubenswrapper[4578]: I1003 13:50:09.434253 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4vlxn"] Oct 03 13:50:09 crc kubenswrapper[4578]: E1003 13:50:09.435414 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="245a6647-a604-425b-8c8d-865f3bfc9fb6" containerName="extract-utilities" Oct 03 13:50:09 crc kubenswrapper[4578]: I1003 13:50:09.435432 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="245a6647-a604-425b-8c8d-865f3bfc9fb6" containerName="extract-utilities" Oct 03 13:50:09 crc kubenswrapper[4578]: E1003 13:50:09.435457 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="245a6647-a604-425b-8c8d-865f3bfc9fb6" containerName="extract-content" Oct 03 13:50:09 crc kubenswrapper[4578]: I1003 13:50:09.435467 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="245a6647-a604-425b-8c8d-865f3bfc9fb6" containerName="extract-content" Oct 03 13:50:09 crc kubenswrapper[4578]: E1003 13:50:09.435486 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="245a6647-a604-425b-8c8d-865f3bfc9fb6" containerName="registry-server" Oct 03 13:50:09 crc kubenswrapper[4578]: I1003 13:50:09.435494 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="245a6647-a604-425b-8c8d-865f3bfc9fb6" containerName="registry-server" Oct 03 13:50:09 crc kubenswrapper[4578]: I1003 13:50:09.436650 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="245a6647-a604-425b-8c8d-865f3bfc9fb6" containerName="registry-server" Oct 03 13:50:09 crc kubenswrapper[4578]: I1003 13:50:09.439143 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4vlxn" Oct 03 13:50:09 crc kubenswrapper[4578]: I1003 13:50:09.448381 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4vlxn"] Oct 03 13:50:09 crc kubenswrapper[4578]: I1003 13:50:09.632358 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbad8b1b-cbde-48b0-a7f6-f8e8494756be-catalog-content\") pod \"redhat-operators-4vlxn\" (UID: \"bbad8b1b-cbde-48b0-a7f6-f8e8494756be\") " pod="openshift-marketplace/redhat-operators-4vlxn" Oct 03 13:50:09 crc kubenswrapper[4578]: I1003 13:50:09.632438 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5w2g\" (UniqueName: \"kubernetes.io/projected/bbad8b1b-cbde-48b0-a7f6-f8e8494756be-kube-api-access-c5w2g\") pod \"redhat-operators-4vlxn\" (UID: \"bbad8b1b-cbde-48b0-a7f6-f8e8494756be\") " pod="openshift-marketplace/redhat-operators-4vlxn" Oct 03 13:50:09 crc kubenswrapper[4578]: I1003 13:50:09.632594 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbad8b1b-cbde-48b0-a7f6-f8e8494756be-utilities\") pod \"redhat-operators-4vlxn\" (UID: \"bbad8b1b-cbde-48b0-a7f6-f8e8494756be\") " pod="openshift-marketplace/redhat-operators-4vlxn" Oct 03 13:50:09 crc kubenswrapper[4578]: I1003 13:50:09.734461 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbad8b1b-cbde-48b0-a7f6-f8e8494756be-utilities\") pod \"redhat-operators-4vlxn\" (UID: \"bbad8b1b-cbde-48b0-a7f6-f8e8494756be\") " pod="openshift-marketplace/redhat-operators-4vlxn" Oct 03 13:50:09 crc kubenswrapper[4578]: I1003 13:50:09.734598 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbad8b1b-cbde-48b0-a7f6-f8e8494756be-catalog-content\") pod \"redhat-operators-4vlxn\" (UID: \"bbad8b1b-cbde-48b0-a7f6-f8e8494756be\") " pod="openshift-marketplace/redhat-operators-4vlxn" Oct 03 13:50:09 crc kubenswrapper[4578]: I1003 13:50:09.734666 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5w2g\" (UniqueName: \"kubernetes.io/projected/bbad8b1b-cbde-48b0-a7f6-f8e8494756be-kube-api-access-c5w2g\") pod \"redhat-operators-4vlxn\" (UID: \"bbad8b1b-cbde-48b0-a7f6-f8e8494756be\") " pod="openshift-marketplace/redhat-operators-4vlxn" Oct 03 13:50:09 crc kubenswrapper[4578]: I1003 13:50:09.734918 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbad8b1b-cbde-48b0-a7f6-f8e8494756be-utilities\") pod \"redhat-operators-4vlxn\" (UID: \"bbad8b1b-cbde-48b0-a7f6-f8e8494756be\") " pod="openshift-marketplace/redhat-operators-4vlxn" Oct 03 13:50:09 crc kubenswrapper[4578]: I1003 13:50:09.735099 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbad8b1b-cbde-48b0-a7f6-f8e8494756be-catalog-content\") pod \"redhat-operators-4vlxn\" (UID: \"bbad8b1b-cbde-48b0-a7f6-f8e8494756be\") " pod="openshift-marketplace/redhat-operators-4vlxn" Oct 03 13:50:09 crc kubenswrapper[4578]: I1003 13:50:09.752602 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5w2g\" (UniqueName: \"kubernetes.io/projected/bbad8b1b-cbde-48b0-a7f6-f8e8494756be-kube-api-access-c5w2g\") pod \"redhat-operators-4vlxn\" (UID: \"bbad8b1b-cbde-48b0-a7f6-f8e8494756be\") " pod="openshift-marketplace/redhat-operators-4vlxn" Oct 03 13:50:09 crc kubenswrapper[4578]: I1003 13:50:09.764502 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4vlxn" Oct 03 13:50:10 crc kubenswrapper[4578]: I1003 13:50:10.238543 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4vlxn"] Oct 03 13:50:10 crc kubenswrapper[4578]: I1003 13:50:10.800859 4578 generic.go:334] "Generic (PLEG): container finished" podID="bbad8b1b-cbde-48b0-a7f6-f8e8494756be" containerID="b619734a3e3899bbcd99f09973f13561a780225f89590c188c0755f1348128ba" exitCode=0 Oct 03 13:50:10 crc kubenswrapper[4578]: I1003 13:50:10.800901 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vlxn" event={"ID":"bbad8b1b-cbde-48b0-a7f6-f8e8494756be","Type":"ContainerDied","Data":"b619734a3e3899bbcd99f09973f13561a780225f89590c188c0755f1348128ba"} Oct 03 13:50:10 crc kubenswrapper[4578]: I1003 13:50:10.800928 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vlxn" event={"ID":"bbad8b1b-cbde-48b0-a7f6-f8e8494756be","Type":"ContainerStarted","Data":"e7efd5d39d2de382b023be0799abd409427ec05ccc98cc038450ba694a1803ed"} Oct 03 13:50:12 crc kubenswrapper[4578]: I1003 13:50:12.819021 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vlxn" event={"ID":"bbad8b1b-cbde-48b0-a7f6-f8e8494756be","Type":"ContainerStarted","Data":"3a5a2cae910cfea4985b96038031498e86f4ee87bf9ae01cfbaf58bfa8f34582"} Oct 03 13:50:17 crc kubenswrapper[4578]: I1003 13:50:17.873348 4578 generic.go:334] "Generic (PLEG): container finished" podID="bbad8b1b-cbde-48b0-a7f6-f8e8494756be" containerID="3a5a2cae910cfea4985b96038031498e86f4ee87bf9ae01cfbaf58bfa8f34582" exitCode=0 Oct 03 13:50:17 crc kubenswrapper[4578]: I1003 13:50:17.873422 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vlxn" event={"ID":"bbad8b1b-cbde-48b0-a7f6-f8e8494756be","Type":"ContainerDied","Data":"3a5a2cae910cfea4985b96038031498e86f4ee87bf9ae01cfbaf58bfa8f34582"} Oct 03 13:50:18 crc kubenswrapper[4578]: I1003 13:50:18.883267 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vlxn" event={"ID":"bbad8b1b-cbde-48b0-a7f6-f8e8494756be","Type":"ContainerStarted","Data":"7efd033a17d6aab3963b0d02607dc45af754c27d6ee1ee0da0048470265e0fcd"} Oct 03 13:50:18 crc kubenswrapper[4578]: I1003 13:50:18.900034 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4vlxn" podStartSLOduration=2.301275014 podStartE2EDuration="9.900008514s" podCreationTimestamp="2025-10-03 13:50:09 +0000 UTC" firstStartedPulling="2025-10-03 13:50:10.803259899 +0000 UTC m=+3546.601732083" lastFinishedPulling="2025-10-03 13:50:18.401993399 +0000 UTC m=+3554.200465583" observedRunningTime="2025-10-03 13:50:18.89893802 +0000 UTC m=+3554.697410214" watchObservedRunningTime="2025-10-03 13:50:18.900008514 +0000 UTC m=+3554.698480698" Oct 03 13:50:19 crc kubenswrapper[4578]: I1003 13:50:19.765309 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4vlxn" Oct 03 13:50:19 crc kubenswrapper[4578]: I1003 13:50:19.765747 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4vlxn" Oct 03 13:50:20 crc kubenswrapper[4578]: I1003 13:50:20.807144 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4vlxn" podUID="bbad8b1b-cbde-48b0-a7f6-f8e8494756be" containerName="registry-server" probeResult="failure" output=< Oct 03 13:50:20 crc kubenswrapper[4578]: timeout: failed to connect service ":50051" within 1s Oct 03 13:50:20 crc kubenswrapper[4578]: > Oct 03 13:50:29 crc kubenswrapper[4578]: I1003 13:50:29.819780 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4vlxn" Oct 03 13:50:29 crc kubenswrapper[4578]: I1003 13:50:29.873913 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4vlxn" Oct 03 13:50:30 crc kubenswrapper[4578]: I1003 13:50:30.066616 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4vlxn"] Oct 03 13:50:30 crc kubenswrapper[4578]: I1003 13:50:30.979518 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4vlxn" podUID="bbad8b1b-cbde-48b0-a7f6-f8e8494756be" containerName="registry-server" containerID="cri-o://7efd033a17d6aab3963b0d02607dc45af754c27d6ee1ee0da0048470265e0fcd" gracePeriod=2 Oct 03 13:50:31 crc kubenswrapper[4578]: I1003 13:50:31.418387 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4vlxn" Oct 03 13:50:31 crc kubenswrapper[4578]: I1003 13:50:31.519447 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbad8b1b-cbde-48b0-a7f6-f8e8494756be-catalog-content\") pod \"bbad8b1b-cbde-48b0-a7f6-f8e8494756be\" (UID: \"bbad8b1b-cbde-48b0-a7f6-f8e8494756be\") " Oct 03 13:50:31 crc kubenswrapper[4578]: I1003 13:50:31.519592 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5w2g\" (UniqueName: \"kubernetes.io/projected/bbad8b1b-cbde-48b0-a7f6-f8e8494756be-kube-api-access-c5w2g\") pod \"bbad8b1b-cbde-48b0-a7f6-f8e8494756be\" (UID: \"bbad8b1b-cbde-48b0-a7f6-f8e8494756be\") " Oct 03 13:50:31 crc kubenswrapper[4578]: I1003 13:50:31.519666 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbad8b1b-cbde-48b0-a7f6-f8e8494756be-utilities\") pod \"bbad8b1b-cbde-48b0-a7f6-f8e8494756be\" (UID: \"bbad8b1b-cbde-48b0-a7f6-f8e8494756be\") " Oct 03 13:50:31 crc kubenswrapper[4578]: I1003 13:50:31.520972 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbad8b1b-cbde-48b0-a7f6-f8e8494756be-utilities" (OuterVolumeSpecName: "utilities") pod "bbad8b1b-cbde-48b0-a7f6-f8e8494756be" (UID: "bbad8b1b-cbde-48b0-a7f6-f8e8494756be"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:50:31 crc kubenswrapper[4578]: I1003 13:50:31.531762 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbad8b1b-cbde-48b0-a7f6-f8e8494756be-kube-api-access-c5w2g" (OuterVolumeSpecName: "kube-api-access-c5w2g") pod "bbad8b1b-cbde-48b0-a7f6-f8e8494756be" (UID: "bbad8b1b-cbde-48b0-a7f6-f8e8494756be"). InnerVolumeSpecName "kube-api-access-c5w2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:50:31 crc kubenswrapper[4578]: I1003 13:50:31.612545 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbad8b1b-cbde-48b0-a7f6-f8e8494756be-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bbad8b1b-cbde-48b0-a7f6-f8e8494756be" (UID: "bbad8b1b-cbde-48b0-a7f6-f8e8494756be"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:50:31 crc kubenswrapper[4578]: I1003 13:50:31.621355 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5w2g\" (UniqueName: \"kubernetes.io/projected/bbad8b1b-cbde-48b0-a7f6-f8e8494756be-kube-api-access-c5w2g\") on node \"crc\" DevicePath \"\"" Oct 03 13:50:31 crc kubenswrapper[4578]: I1003 13:50:31.621392 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbad8b1b-cbde-48b0-a7f6-f8e8494756be-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:50:31 crc kubenswrapper[4578]: I1003 13:50:31.621402 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbad8b1b-cbde-48b0-a7f6-f8e8494756be-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:50:31 crc kubenswrapper[4578]: I1003 13:50:31.988575 4578 generic.go:334] "Generic (PLEG): container finished" podID="bbad8b1b-cbde-48b0-a7f6-f8e8494756be" containerID="7efd033a17d6aab3963b0d02607dc45af754c27d6ee1ee0da0048470265e0fcd" exitCode=0 Oct 03 13:50:31 crc kubenswrapper[4578]: I1003 13:50:31.988671 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4vlxn" Oct 03 13:50:31 crc kubenswrapper[4578]: I1003 13:50:31.988666 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vlxn" event={"ID":"bbad8b1b-cbde-48b0-a7f6-f8e8494756be","Type":"ContainerDied","Data":"7efd033a17d6aab3963b0d02607dc45af754c27d6ee1ee0da0048470265e0fcd"} Oct 03 13:50:31 crc kubenswrapper[4578]: I1003 13:50:31.989165 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vlxn" event={"ID":"bbad8b1b-cbde-48b0-a7f6-f8e8494756be","Type":"ContainerDied","Data":"e7efd5d39d2de382b023be0799abd409427ec05ccc98cc038450ba694a1803ed"} Oct 03 13:50:31 crc kubenswrapper[4578]: I1003 13:50:31.989192 4578 scope.go:117] "RemoveContainer" containerID="7efd033a17d6aab3963b0d02607dc45af754c27d6ee1ee0da0048470265e0fcd" Oct 03 13:50:32 crc kubenswrapper[4578]: I1003 13:50:32.019335 4578 scope.go:117] "RemoveContainer" containerID="3a5a2cae910cfea4985b96038031498e86f4ee87bf9ae01cfbaf58bfa8f34582" Oct 03 13:50:32 crc kubenswrapper[4578]: I1003 13:50:32.024663 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4vlxn"] Oct 03 13:50:32 crc kubenswrapper[4578]: I1003 13:50:32.033940 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4vlxn"] Oct 03 13:50:32 crc kubenswrapper[4578]: I1003 13:50:32.039266 4578 scope.go:117] "RemoveContainer" containerID="b619734a3e3899bbcd99f09973f13561a780225f89590c188c0755f1348128ba" Oct 03 13:50:32 crc kubenswrapper[4578]: I1003 13:50:32.078676 4578 scope.go:117] "RemoveContainer" containerID="7efd033a17d6aab3963b0d02607dc45af754c27d6ee1ee0da0048470265e0fcd" Oct 03 13:50:32 crc kubenswrapper[4578]: E1003 13:50:32.079224 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7efd033a17d6aab3963b0d02607dc45af754c27d6ee1ee0da0048470265e0fcd\": container with ID starting with 7efd033a17d6aab3963b0d02607dc45af754c27d6ee1ee0da0048470265e0fcd not found: ID does not exist" containerID="7efd033a17d6aab3963b0d02607dc45af754c27d6ee1ee0da0048470265e0fcd" Oct 03 13:50:32 crc kubenswrapper[4578]: I1003 13:50:32.079374 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7efd033a17d6aab3963b0d02607dc45af754c27d6ee1ee0da0048470265e0fcd"} err="failed to get container status \"7efd033a17d6aab3963b0d02607dc45af754c27d6ee1ee0da0048470265e0fcd\": rpc error: code = NotFound desc = could not find container \"7efd033a17d6aab3963b0d02607dc45af754c27d6ee1ee0da0048470265e0fcd\": container with ID starting with 7efd033a17d6aab3963b0d02607dc45af754c27d6ee1ee0da0048470265e0fcd not found: ID does not exist" Oct 03 13:50:32 crc kubenswrapper[4578]: I1003 13:50:32.079428 4578 scope.go:117] "RemoveContainer" containerID="3a5a2cae910cfea4985b96038031498e86f4ee87bf9ae01cfbaf58bfa8f34582" Oct 03 13:50:32 crc kubenswrapper[4578]: E1003 13:50:32.079805 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a5a2cae910cfea4985b96038031498e86f4ee87bf9ae01cfbaf58bfa8f34582\": container with ID starting with 3a5a2cae910cfea4985b96038031498e86f4ee87bf9ae01cfbaf58bfa8f34582 not found: ID does not exist" containerID="3a5a2cae910cfea4985b96038031498e86f4ee87bf9ae01cfbaf58bfa8f34582" Oct 03 13:50:32 crc kubenswrapper[4578]: I1003 13:50:32.079844 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a5a2cae910cfea4985b96038031498e86f4ee87bf9ae01cfbaf58bfa8f34582"} err="failed to get container status \"3a5a2cae910cfea4985b96038031498e86f4ee87bf9ae01cfbaf58bfa8f34582\": rpc error: code = NotFound desc = could not find container \"3a5a2cae910cfea4985b96038031498e86f4ee87bf9ae01cfbaf58bfa8f34582\": container with ID starting with 3a5a2cae910cfea4985b96038031498e86f4ee87bf9ae01cfbaf58bfa8f34582 not found: ID does not exist" Oct 03 13:50:32 crc kubenswrapper[4578]: I1003 13:50:32.079870 4578 scope.go:117] "RemoveContainer" containerID="b619734a3e3899bbcd99f09973f13561a780225f89590c188c0755f1348128ba" Oct 03 13:50:32 crc kubenswrapper[4578]: E1003 13:50:32.080256 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b619734a3e3899bbcd99f09973f13561a780225f89590c188c0755f1348128ba\": container with ID starting with b619734a3e3899bbcd99f09973f13561a780225f89590c188c0755f1348128ba not found: ID does not exist" containerID="b619734a3e3899bbcd99f09973f13561a780225f89590c188c0755f1348128ba" Oct 03 13:50:32 crc kubenswrapper[4578]: I1003 13:50:32.080367 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b619734a3e3899bbcd99f09973f13561a780225f89590c188c0755f1348128ba"} err="failed to get container status \"b619734a3e3899bbcd99f09973f13561a780225f89590c188c0755f1348128ba\": rpc error: code = NotFound desc = could not find container \"b619734a3e3899bbcd99f09973f13561a780225f89590c188c0755f1348128ba\": container with ID starting with b619734a3e3899bbcd99f09973f13561a780225f89590c188c0755f1348128ba not found: ID does not exist" Oct 03 13:50:32 crc kubenswrapper[4578]: I1003 13:50:32.919683 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbad8b1b-cbde-48b0-a7f6-f8e8494756be" path="/var/lib/kubelet/pods/bbad8b1b-cbde-48b0-a7f6-f8e8494756be/volumes" Oct 03 13:50:35 crc kubenswrapper[4578]: I1003 13:50:35.091624 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:50:35 crc kubenswrapper[4578]: I1003 13:50:35.092135 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:50:35 crc kubenswrapper[4578]: I1003 13:50:35.092185 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 13:50:35 crc kubenswrapper[4578]: I1003 13:50:35.093024 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e"} pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:50:35 crc kubenswrapper[4578]: I1003 13:50:35.093081 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" containerID="cri-o://ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" gracePeriod=600 Oct 03 13:50:35 crc kubenswrapper[4578]: E1003 13:50:35.213245 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:50:36 crc kubenswrapper[4578]: I1003 13:50:36.032762 4578 generic.go:334] "Generic (PLEG): container finished" podID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" exitCode=0 Oct 03 13:50:36 crc kubenswrapper[4578]: I1003 13:50:36.032885 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerDied","Data":"ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e"} Oct 03 13:50:36 crc kubenswrapper[4578]: I1003 13:50:36.033249 4578 scope.go:117] "RemoveContainer" containerID="63284c2a61274b6348f80287108612ab5c7b2a261fcd4d7590b033486c46447c" Oct 03 13:50:36 crc kubenswrapper[4578]: I1003 13:50:36.034529 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:50:36 crc kubenswrapper[4578]: E1003 13:50:36.034883 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:50:50 crc kubenswrapper[4578]: I1003 13:50:50.909715 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:50:50 crc kubenswrapper[4578]: E1003 13:50:50.910773 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:51:02 crc kubenswrapper[4578]: I1003 13:51:02.909937 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:51:02 crc kubenswrapper[4578]: E1003 13:51:02.911097 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:51:14 crc kubenswrapper[4578]: I1003 13:51:14.915778 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:51:14 crc kubenswrapper[4578]: E1003 13:51:14.916708 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:51:29 crc kubenswrapper[4578]: I1003 13:51:29.909954 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:51:29 crc kubenswrapper[4578]: E1003 13:51:29.910913 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:51:33 crc kubenswrapper[4578]: I1003 13:51:33.713309 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fj25z"] Oct 03 13:51:33 crc kubenswrapper[4578]: E1003 13:51:33.714439 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbad8b1b-cbde-48b0-a7f6-f8e8494756be" containerName="extract-content" Oct 03 13:51:33 crc kubenswrapper[4578]: I1003 13:51:33.714454 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbad8b1b-cbde-48b0-a7f6-f8e8494756be" containerName="extract-content" Oct 03 13:51:33 crc kubenswrapper[4578]: E1003 13:51:33.714477 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbad8b1b-cbde-48b0-a7f6-f8e8494756be" containerName="registry-server" Oct 03 13:51:33 crc kubenswrapper[4578]: I1003 13:51:33.714483 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbad8b1b-cbde-48b0-a7f6-f8e8494756be" containerName="registry-server" Oct 03 13:51:33 crc kubenswrapper[4578]: E1003 13:51:33.714514 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbad8b1b-cbde-48b0-a7f6-f8e8494756be" containerName="extract-utilities" Oct 03 13:51:33 crc kubenswrapper[4578]: I1003 13:51:33.714522 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbad8b1b-cbde-48b0-a7f6-f8e8494756be" containerName="extract-utilities" Oct 03 13:51:33 crc kubenswrapper[4578]: I1003 13:51:33.714709 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbad8b1b-cbde-48b0-a7f6-f8e8494756be" containerName="registry-server" Oct 03 13:51:33 crc kubenswrapper[4578]: I1003 13:51:33.715951 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fj25z" Oct 03 13:51:33 crc kubenswrapper[4578]: I1003 13:51:33.732211 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fj25z"] Oct 03 13:51:33 crc kubenswrapper[4578]: I1003 13:51:33.823140 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6dbda96-8e27-4135-a771-2ec8e4f15b19-catalog-content\") pod \"redhat-marketplace-fj25z\" (UID: \"a6dbda96-8e27-4135-a771-2ec8e4f15b19\") " pod="openshift-marketplace/redhat-marketplace-fj25z" Oct 03 13:51:33 crc kubenswrapper[4578]: I1003 13:51:33.823178 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg54s\" (UniqueName: \"kubernetes.io/projected/a6dbda96-8e27-4135-a771-2ec8e4f15b19-kube-api-access-cg54s\") pod \"redhat-marketplace-fj25z\" (UID: \"a6dbda96-8e27-4135-a771-2ec8e4f15b19\") " pod="openshift-marketplace/redhat-marketplace-fj25z" Oct 03 13:51:33 crc kubenswrapper[4578]: I1003 13:51:33.823214 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6dbda96-8e27-4135-a771-2ec8e4f15b19-utilities\") pod \"redhat-marketplace-fj25z\" (UID: \"a6dbda96-8e27-4135-a771-2ec8e4f15b19\") " pod="openshift-marketplace/redhat-marketplace-fj25z" Oct 03 13:51:33 crc kubenswrapper[4578]: I1003 13:51:33.924875 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6dbda96-8e27-4135-a771-2ec8e4f15b19-catalog-content\") pod \"redhat-marketplace-fj25z\" (UID: \"a6dbda96-8e27-4135-a771-2ec8e4f15b19\") " pod="openshift-marketplace/redhat-marketplace-fj25z" Oct 03 13:51:33 crc kubenswrapper[4578]: I1003 13:51:33.924921 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg54s\" (UniqueName: \"kubernetes.io/projected/a6dbda96-8e27-4135-a771-2ec8e4f15b19-kube-api-access-cg54s\") pod \"redhat-marketplace-fj25z\" (UID: \"a6dbda96-8e27-4135-a771-2ec8e4f15b19\") " pod="openshift-marketplace/redhat-marketplace-fj25z" Oct 03 13:51:33 crc kubenswrapper[4578]: I1003 13:51:33.924969 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6dbda96-8e27-4135-a771-2ec8e4f15b19-utilities\") pod \"redhat-marketplace-fj25z\" (UID: \"a6dbda96-8e27-4135-a771-2ec8e4f15b19\") " pod="openshift-marketplace/redhat-marketplace-fj25z" Oct 03 13:51:33 crc kubenswrapper[4578]: I1003 13:51:33.929094 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6dbda96-8e27-4135-a771-2ec8e4f15b19-catalog-content\") pod \"redhat-marketplace-fj25z\" (UID: \"a6dbda96-8e27-4135-a771-2ec8e4f15b19\") " pod="openshift-marketplace/redhat-marketplace-fj25z" Oct 03 13:51:33 crc kubenswrapper[4578]: I1003 13:51:33.929190 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6dbda96-8e27-4135-a771-2ec8e4f15b19-utilities\") pod \"redhat-marketplace-fj25z\" (UID: \"a6dbda96-8e27-4135-a771-2ec8e4f15b19\") " pod="openshift-marketplace/redhat-marketplace-fj25z" Oct 03 13:51:33 crc kubenswrapper[4578]: I1003 13:51:33.964123 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg54s\" (UniqueName: \"kubernetes.io/projected/a6dbda96-8e27-4135-a771-2ec8e4f15b19-kube-api-access-cg54s\") pod \"redhat-marketplace-fj25z\" (UID: \"a6dbda96-8e27-4135-a771-2ec8e4f15b19\") " pod="openshift-marketplace/redhat-marketplace-fj25z" Oct 03 13:51:34 crc kubenswrapper[4578]: I1003 13:51:34.042210 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fj25z" Oct 03 13:51:34 crc kubenswrapper[4578]: I1003 13:51:34.499354 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fj25z"] Oct 03 13:51:34 crc kubenswrapper[4578]: I1003 13:51:34.565616 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fj25z" event={"ID":"a6dbda96-8e27-4135-a771-2ec8e4f15b19","Type":"ContainerStarted","Data":"8a9499a54f9033a79510c867cb84f3b13a0318f92e94558c043f8f13631aedb2"} Oct 03 13:51:35 crc kubenswrapper[4578]: I1003 13:51:35.575065 4578 generic.go:334] "Generic (PLEG): container finished" podID="a6dbda96-8e27-4135-a771-2ec8e4f15b19" containerID="b23f6726a39f235ebfaa58679de5b69996c95ea19ab7f600d7fd05674a78be37" exitCode=0 Oct 03 13:51:35 crc kubenswrapper[4578]: I1003 13:51:35.575392 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fj25z" event={"ID":"a6dbda96-8e27-4135-a771-2ec8e4f15b19","Type":"ContainerDied","Data":"b23f6726a39f235ebfaa58679de5b69996c95ea19ab7f600d7fd05674a78be37"} Oct 03 13:51:36 crc kubenswrapper[4578]: I1003 13:51:36.584135 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fj25z" event={"ID":"a6dbda96-8e27-4135-a771-2ec8e4f15b19","Type":"ContainerStarted","Data":"d642e9f2bdd9853488b17ae72bc7a400a8fb48d04494dcfeb47cfaa89b445720"} Oct 03 13:51:37 crc kubenswrapper[4578]: I1003 13:51:37.602008 4578 generic.go:334] "Generic (PLEG): container finished" podID="a6dbda96-8e27-4135-a771-2ec8e4f15b19" containerID="d642e9f2bdd9853488b17ae72bc7a400a8fb48d04494dcfeb47cfaa89b445720" exitCode=0 Oct 03 13:51:37 crc kubenswrapper[4578]: I1003 13:51:37.602059 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fj25z" event={"ID":"a6dbda96-8e27-4135-a771-2ec8e4f15b19","Type":"ContainerDied","Data":"d642e9f2bdd9853488b17ae72bc7a400a8fb48d04494dcfeb47cfaa89b445720"} Oct 03 13:51:38 crc kubenswrapper[4578]: I1003 13:51:38.612219 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fj25z" event={"ID":"a6dbda96-8e27-4135-a771-2ec8e4f15b19","Type":"ContainerStarted","Data":"4a570276e701b4a0231dcf5fb6873aaab9eb09040c9e3c1b540b10fa9a08d702"} Oct 03 13:51:38 crc kubenswrapper[4578]: I1003 13:51:38.631840 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fj25z" podStartSLOduration=3.068251782 podStartE2EDuration="5.631821415s" podCreationTimestamp="2025-10-03 13:51:33 +0000 UTC" firstStartedPulling="2025-10-03 13:51:35.577740472 +0000 UTC m=+3631.376212666" lastFinishedPulling="2025-10-03 13:51:38.141310115 +0000 UTC m=+3633.939782299" observedRunningTime="2025-10-03 13:51:38.62908763 +0000 UTC m=+3634.427559824" watchObservedRunningTime="2025-10-03 13:51:38.631821415 +0000 UTC m=+3634.430293599" Oct 03 13:51:43 crc kubenswrapper[4578]: I1003 13:51:43.909722 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:51:43 crc kubenswrapper[4578]: E1003 13:51:43.910596 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:51:44 crc kubenswrapper[4578]: I1003 13:51:44.043024 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fj25z" Oct 03 13:51:44 crc kubenswrapper[4578]: I1003 13:51:44.043083 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fj25z" Oct 03 13:51:44 crc kubenswrapper[4578]: I1003 13:51:44.084215 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fj25z" Oct 03 13:51:44 crc kubenswrapper[4578]: I1003 13:51:44.709306 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fj25z" Oct 03 13:51:44 crc kubenswrapper[4578]: I1003 13:51:44.769490 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fj25z"] Oct 03 13:51:46 crc kubenswrapper[4578]: I1003 13:51:46.673350 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fj25z" podUID="a6dbda96-8e27-4135-a771-2ec8e4f15b19" containerName="registry-server" containerID="cri-o://4a570276e701b4a0231dcf5fb6873aaab9eb09040c9e3c1b540b10fa9a08d702" gracePeriod=2 Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.159370 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fj25z" Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.299990 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6dbda96-8e27-4135-a771-2ec8e4f15b19-catalog-content\") pod \"a6dbda96-8e27-4135-a771-2ec8e4f15b19\" (UID: \"a6dbda96-8e27-4135-a771-2ec8e4f15b19\") " Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.300132 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6dbda96-8e27-4135-a771-2ec8e4f15b19-utilities\") pod \"a6dbda96-8e27-4135-a771-2ec8e4f15b19\" (UID: \"a6dbda96-8e27-4135-a771-2ec8e4f15b19\") " Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.300193 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cg54s\" (UniqueName: \"kubernetes.io/projected/a6dbda96-8e27-4135-a771-2ec8e4f15b19-kube-api-access-cg54s\") pod \"a6dbda96-8e27-4135-a771-2ec8e4f15b19\" (UID: \"a6dbda96-8e27-4135-a771-2ec8e4f15b19\") " Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.301891 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6dbda96-8e27-4135-a771-2ec8e4f15b19-utilities" (OuterVolumeSpecName: "utilities") pod "a6dbda96-8e27-4135-a771-2ec8e4f15b19" (UID: "a6dbda96-8e27-4135-a771-2ec8e4f15b19"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.308886 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6dbda96-8e27-4135-a771-2ec8e4f15b19-kube-api-access-cg54s" (OuterVolumeSpecName: "kube-api-access-cg54s") pod "a6dbda96-8e27-4135-a771-2ec8e4f15b19" (UID: "a6dbda96-8e27-4135-a771-2ec8e4f15b19"). InnerVolumeSpecName "kube-api-access-cg54s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.317577 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6dbda96-8e27-4135-a771-2ec8e4f15b19-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a6dbda96-8e27-4135-a771-2ec8e4f15b19" (UID: "a6dbda96-8e27-4135-a771-2ec8e4f15b19"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.403062 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cg54s\" (UniqueName: \"kubernetes.io/projected/a6dbda96-8e27-4135-a771-2ec8e4f15b19-kube-api-access-cg54s\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.403104 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6dbda96-8e27-4135-a771-2ec8e4f15b19-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.403117 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6dbda96-8e27-4135-a771-2ec8e4f15b19-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.685860 4578 generic.go:334] "Generic (PLEG): container finished" podID="a6dbda96-8e27-4135-a771-2ec8e4f15b19" containerID="4a570276e701b4a0231dcf5fb6873aaab9eb09040c9e3c1b540b10fa9a08d702" exitCode=0 Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.686037 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fj25z" event={"ID":"a6dbda96-8e27-4135-a771-2ec8e4f15b19","Type":"ContainerDied","Data":"4a570276e701b4a0231dcf5fb6873aaab9eb09040c9e3c1b540b10fa9a08d702"} Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.686202 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fj25z" event={"ID":"a6dbda96-8e27-4135-a771-2ec8e4f15b19","Type":"ContainerDied","Data":"8a9499a54f9033a79510c867cb84f3b13a0318f92e94558c043f8f13631aedb2"} Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.686227 4578 scope.go:117] "RemoveContainer" containerID="4a570276e701b4a0231dcf5fb6873aaab9eb09040c9e3c1b540b10fa9a08d702" Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.686146 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fj25z" Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.717452 4578 scope.go:117] "RemoveContainer" containerID="d642e9f2bdd9853488b17ae72bc7a400a8fb48d04494dcfeb47cfaa89b445720" Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.723030 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fj25z"] Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.733410 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fj25z"] Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.747591 4578 scope.go:117] "RemoveContainer" containerID="b23f6726a39f235ebfaa58679de5b69996c95ea19ab7f600d7fd05674a78be37" Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.807836 4578 scope.go:117] "RemoveContainer" containerID="4a570276e701b4a0231dcf5fb6873aaab9eb09040c9e3c1b540b10fa9a08d702" Oct 03 13:51:47 crc kubenswrapper[4578]: E1003 13:51:47.808365 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a570276e701b4a0231dcf5fb6873aaab9eb09040c9e3c1b540b10fa9a08d702\": container with ID starting with 4a570276e701b4a0231dcf5fb6873aaab9eb09040c9e3c1b540b10fa9a08d702 not found: ID does not exist" containerID="4a570276e701b4a0231dcf5fb6873aaab9eb09040c9e3c1b540b10fa9a08d702" Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.808416 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a570276e701b4a0231dcf5fb6873aaab9eb09040c9e3c1b540b10fa9a08d702"} err="failed to get container status \"4a570276e701b4a0231dcf5fb6873aaab9eb09040c9e3c1b540b10fa9a08d702\": rpc error: code = NotFound desc = could not find container \"4a570276e701b4a0231dcf5fb6873aaab9eb09040c9e3c1b540b10fa9a08d702\": container with ID starting with 4a570276e701b4a0231dcf5fb6873aaab9eb09040c9e3c1b540b10fa9a08d702 not found: ID does not exist" Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.808445 4578 scope.go:117] "RemoveContainer" containerID="d642e9f2bdd9853488b17ae72bc7a400a8fb48d04494dcfeb47cfaa89b445720" Oct 03 13:51:47 crc kubenswrapper[4578]: E1003 13:51:47.808870 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d642e9f2bdd9853488b17ae72bc7a400a8fb48d04494dcfeb47cfaa89b445720\": container with ID starting with d642e9f2bdd9853488b17ae72bc7a400a8fb48d04494dcfeb47cfaa89b445720 not found: ID does not exist" containerID="d642e9f2bdd9853488b17ae72bc7a400a8fb48d04494dcfeb47cfaa89b445720" Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.808911 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d642e9f2bdd9853488b17ae72bc7a400a8fb48d04494dcfeb47cfaa89b445720"} err="failed to get container status \"d642e9f2bdd9853488b17ae72bc7a400a8fb48d04494dcfeb47cfaa89b445720\": rpc error: code = NotFound desc = could not find container \"d642e9f2bdd9853488b17ae72bc7a400a8fb48d04494dcfeb47cfaa89b445720\": container with ID starting with d642e9f2bdd9853488b17ae72bc7a400a8fb48d04494dcfeb47cfaa89b445720 not found: ID does not exist" Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.808938 4578 scope.go:117] "RemoveContainer" containerID="b23f6726a39f235ebfaa58679de5b69996c95ea19ab7f600d7fd05674a78be37" Oct 03 13:51:47 crc kubenswrapper[4578]: E1003 13:51:47.809272 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b23f6726a39f235ebfaa58679de5b69996c95ea19ab7f600d7fd05674a78be37\": container with ID starting with b23f6726a39f235ebfaa58679de5b69996c95ea19ab7f600d7fd05674a78be37 not found: ID does not exist" containerID="b23f6726a39f235ebfaa58679de5b69996c95ea19ab7f600d7fd05674a78be37" Oct 03 13:51:47 crc kubenswrapper[4578]: I1003 13:51:47.809298 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b23f6726a39f235ebfaa58679de5b69996c95ea19ab7f600d7fd05674a78be37"} err="failed to get container status \"b23f6726a39f235ebfaa58679de5b69996c95ea19ab7f600d7fd05674a78be37\": rpc error: code = NotFound desc = could not find container \"b23f6726a39f235ebfaa58679de5b69996c95ea19ab7f600d7fd05674a78be37\": container with ID starting with b23f6726a39f235ebfaa58679de5b69996c95ea19ab7f600d7fd05674a78be37 not found: ID does not exist" Oct 03 13:51:48 crc kubenswrapper[4578]: I1003 13:51:48.920157 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6dbda96-8e27-4135-a771-2ec8e4f15b19" path="/var/lib/kubelet/pods/a6dbda96-8e27-4135-a771-2ec8e4f15b19/volumes" Oct 03 13:51:56 crc kubenswrapper[4578]: I1003 13:51:56.909674 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:51:56 crc kubenswrapper[4578]: E1003 13:51:56.911237 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:52:10 crc kubenswrapper[4578]: I1003 13:52:10.915485 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:52:10 crc kubenswrapper[4578]: E1003 13:52:10.917118 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:52:22 crc kubenswrapper[4578]: I1003 13:52:22.909458 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:52:22 crc kubenswrapper[4578]: E1003 13:52:22.910292 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:52:34 crc kubenswrapper[4578]: I1003 13:52:34.915239 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:52:34 crc kubenswrapper[4578]: E1003 13:52:34.916034 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:52:49 crc kubenswrapper[4578]: I1003 13:52:49.909837 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:52:49 crc kubenswrapper[4578]: E1003 13:52:49.910646 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:53:04 crc kubenswrapper[4578]: I1003 13:53:04.916048 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:53:04 crc kubenswrapper[4578]: E1003 13:53:04.916872 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:53:19 crc kubenswrapper[4578]: I1003 13:53:19.908889 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:53:19 crc kubenswrapper[4578]: E1003 13:53:19.909817 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:53:32 crc kubenswrapper[4578]: I1003 13:53:32.909241 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:53:32 crc kubenswrapper[4578]: E1003 13:53:32.910198 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:53:43 crc kubenswrapper[4578]: I1003 13:53:43.908997 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:53:43 crc kubenswrapper[4578]: E1003 13:53:43.909783 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:53:57 crc kubenswrapper[4578]: I1003 13:53:57.909228 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:53:57 crc kubenswrapper[4578]: E1003 13:53:57.910198 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:54:05 crc kubenswrapper[4578]: I1003 13:54:05.979005 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sv9qf"] Oct 03 13:54:05 crc kubenswrapper[4578]: E1003 13:54:05.980143 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6dbda96-8e27-4135-a771-2ec8e4f15b19" containerName="extract-utilities" Oct 03 13:54:05 crc kubenswrapper[4578]: I1003 13:54:05.980162 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6dbda96-8e27-4135-a771-2ec8e4f15b19" containerName="extract-utilities" Oct 03 13:54:05 crc kubenswrapper[4578]: E1003 13:54:05.980207 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6dbda96-8e27-4135-a771-2ec8e4f15b19" containerName="registry-server" Oct 03 13:54:05 crc kubenswrapper[4578]: I1003 13:54:05.980242 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6dbda96-8e27-4135-a771-2ec8e4f15b19" containerName="registry-server" Oct 03 13:54:05 crc kubenswrapper[4578]: E1003 13:54:05.980254 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6dbda96-8e27-4135-a771-2ec8e4f15b19" containerName="extract-content" Oct 03 13:54:05 crc kubenswrapper[4578]: I1003 13:54:05.980262 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6dbda96-8e27-4135-a771-2ec8e4f15b19" containerName="extract-content" Oct 03 13:54:05 crc kubenswrapper[4578]: I1003 13:54:05.980485 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6dbda96-8e27-4135-a771-2ec8e4f15b19" containerName="registry-server" Oct 03 13:54:05 crc kubenswrapper[4578]: I1003 13:54:05.981869 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sv9qf" Oct 03 13:54:06 crc kubenswrapper[4578]: I1003 13:54:06.004565 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sv9qf"] Oct 03 13:54:06 crc kubenswrapper[4578]: I1003 13:54:06.120121 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlj2w\" (UniqueName: \"kubernetes.io/projected/94d1860f-c6d0-4d65-b7b7-c70b945d9211-kube-api-access-jlj2w\") pod \"community-operators-sv9qf\" (UID: \"94d1860f-c6d0-4d65-b7b7-c70b945d9211\") " pod="openshift-marketplace/community-operators-sv9qf" Oct 03 13:54:06 crc kubenswrapper[4578]: I1003 13:54:06.120235 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94d1860f-c6d0-4d65-b7b7-c70b945d9211-utilities\") pod \"community-operators-sv9qf\" (UID: \"94d1860f-c6d0-4d65-b7b7-c70b945d9211\") " pod="openshift-marketplace/community-operators-sv9qf" Oct 03 13:54:06 crc kubenswrapper[4578]: I1003 13:54:06.120298 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94d1860f-c6d0-4d65-b7b7-c70b945d9211-catalog-content\") pod \"community-operators-sv9qf\" (UID: \"94d1860f-c6d0-4d65-b7b7-c70b945d9211\") " pod="openshift-marketplace/community-operators-sv9qf" Oct 03 13:54:06 crc kubenswrapper[4578]: I1003 13:54:06.222024 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94d1860f-c6d0-4d65-b7b7-c70b945d9211-utilities\") pod \"community-operators-sv9qf\" (UID: \"94d1860f-c6d0-4d65-b7b7-c70b945d9211\") " pod="openshift-marketplace/community-operators-sv9qf" Oct 03 13:54:06 crc kubenswrapper[4578]: I1003 13:54:06.222106 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94d1860f-c6d0-4d65-b7b7-c70b945d9211-catalog-content\") pod \"community-operators-sv9qf\" (UID: \"94d1860f-c6d0-4d65-b7b7-c70b945d9211\") " pod="openshift-marketplace/community-operators-sv9qf" Oct 03 13:54:06 crc kubenswrapper[4578]: I1003 13:54:06.222173 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlj2w\" (UniqueName: \"kubernetes.io/projected/94d1860f-c6d0-4d65-b7b7-c70b945d9211-kube-api-access-jlj2w\") pod \"community-operators-sv9qf\" (UID: \"94d1860f-c6d0-4d65-b7b7-c70b945d9211\") " pod="openshift-marketplace/community-operators-sv9qf" Oct 03 13:54:06 crc kubenswrapper[4578]: I1003 13:54:06.222570 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94d1860f-c6d0-4d65-b7b7-c70b945d9211-utilities\") pod \"community-operators-sv9qf\" (UID: \"94d1860f-c6d0-4d65-b7b7-c70b945d9211\") " pod="openshift-marketplace/community-operators-sv9qf" Oct 03 13:54:06 crc kubenswrapper[4578]: I1003 13:54:06.222602 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94d1860f-c6d0-4d65-b7b7-c70b945d9211-catalog-content\") pod \"community-operators-sv9qf\" (UID: \"94d1860f-c6d0-4d65-b7b7-c70b945d9211\") " pod="openshift-marketplace/community-operators-sv9qf" Oct 03 13:54:06 crc kubenswrapper[4578]: I1003 13:54:06.257814 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlj2w\" (UniqueName: \"kubernetes.io/projected/94d1860f-c6d0-4d65-b7b7-c70b945d9211-kube-api-access-jlj2w\") pod \"community-operators-sv9qf\" (UID: \"94d1860f-c6d0-4d65-b7b7-c70b945d9211\") " pod="openshift-marketplace/community-operators-sv9qf" Oct 03 13:54:06 crc kubenswrapper[4578]: I1003 13:54:06.313037 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sv9qf" Oct 03 13:54:06 crc kubenswrapper[4578]: I1003 13:54:06.873414 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sv9qf"] Oct 03 13:54:06 crc kubenswrapper[4578]: I1003 13:54:06.918710 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sv9qf" event={"ID":"94d1860f-c6d0-4d65-b7b7-c70b945d9211","Type":"ContainerStarted","Data":"de62b087c08995d678b2e9d8ba6c3e9a1bd311b6b8613843f647561bfa816872"} Oct 03 13:54:07 crc kubenswrapper[4578]: I1003 13:54:07.928534 4578 generic.go:334] "Generic (PLEG): container finished" podID="94d1860f-c6d0-4d65-b7b7-c70b945d9211" containerID="8de6621d073ab8ec23594007c4317571e435e0b9750ed78498c1a2aec8b87c51" exitCode=0 Oct 03 13:54:07 crc kubenswrapper[4578]: I1003 13:54:07.928874 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sv9qf" event={"ID":"94d1860f-c6d0-4d65-b7b7-c70b945d9211","Type":"ContainerDied","Data":"8de6621d073ab8ec23594007c4317571e435e0b9750ed78498c1a2aec8b87c51"} Oct 03 13:54:07 crc kubenswrapper[4578]: I1003 13:54:07.931231 4578 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 13:54:09 crc kubenswrapper[4578]: I1003 13:54:09.954851 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sv9qf" event={"ID":"94d1860f-c6d0-4d65-b7b7-c70b945d9211","Type":"ContainerStarted","Data":"6971daa88c697335ce3a470e5c72f2a7ee7f2073e99d21b81e7f0d4791f2a052"} Oct 03 13:54:10 crc kubenswrapper[4578]: I1003 13:54:10.909808 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:54:10 crc kubenswrapper[4578]: E1003 13:54:10.910542 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:54:10 crc kubenswrapper[4578]: I1003 13:54:10.973688 4578 generic.go:334] "Generic (PLEG): container finished" podID="94d1860f-c6d0-4d65-b7b7-c70b945d9211" containerID="6971daa88c697335ce3a470e5c72f2a7ee7f2073e99d21b81e7f0d4791f2a052" exitCode=0 Oct 03 13:54:10 crc kubenswrapper[4578]: I1003 13:54:10.973742 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sv9qf" event={"ID":"94d1860f-c6d0-4d65-b7b7-c70b945d9211","Type":"ContainerDied","Data":"6971daa88c697335ce3a470e5c72f2a7ee7f2073e99d21b81e7f0d4791f2a052"} Oct 03 13:54:11 crc kubenswrapper[4578]: I1003 13:54:11.985909 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sv9qf" event={"ID":"94d1860f-c6d0-4d65-b7b7-c70b945d9211","Type":"ContainerStarted","Data":"f47c30e91474c3a53994aca525e4e26a8a07f910737215f83b86c3b04617c359"} Oct 03 13:54:12 crc kubenswrapper[4578]: I1003 13:54:12.007474 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sv9qf" podStartSLOduration=3.5372886770000003 podStartE2EDuration="7.007455258s" podCreationTimestamp="2025-10-03 13:54:05 +0000 UTC" firstStartedPulling="2025-10-03 13:54:07.931034206 +0000 UTC m=+3783.729506390" lastFinishedPulling="2025-10-03 13:54:11.401200787 +0000 UTC m=+3787.199672971" observedRunningTime="2025-10-03 13:54:12.000803791 +0000 UTC m=+3787.799275975" watchObservedRunningTime="2025-10-03 13:54:12.007455258 +0000 UTC m=+3787.805927442" Oct 03 13:54:16 crc kubenswrapper[4578]: I1003 13:54:16.314462 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sv9qf" Oct 03 13:54:16 crc kubenswrapper[4578]: I1003 13:54:16.315034 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sv9qf" Oct 03 13:54:16 crc kubenswrapper[4578]: I1003 13:54:16.374461 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sv9qf" Oct 03 13:54:17 crc kubenswrapper[4578]: I1003 13:54:17.074991 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sv9qf" Oct 03 13:54:17 crc kubenswrapper[4578]: I1003 13:54:17.133797 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sv9qf"] Oct 03 13:54:19 crc kubenswrapper[4578]: I1003 13:54:19.043464 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sv9qf" podUID="94d1860f-c6d0-4d65-b7b7-c70b945d9211" containerName="registry-server" containerID="cri-o://f47c30e91474c3a53994aca525e4e26a8a07f910737215f83b86c3b04617c359" gracePeriod=2 Oct 03 13:54:19 crc kubenswrapper[4578]: I1003 13:54:19.536607 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sv9qf" Oct 03 13:54:19 crc kubenswrapper[4578]: I1003 13:54:19.682606 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94d1860f-c6d0-4d65-b7b7-c70b945d9211-utilities\") pod \"94d1860f-c6d0-4d65-b7b7-c70b945d9211\" (UID: \"94d1860f-c6d0-4d65-b7b7-c70b945d9211\") " Oct 03 13:54:19 crc kubenswrapper[4578]: I1003 13:54:19.682780 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94d1860f-c6d0-4d65-b7b7-c70b945d9211-catalog-content\") pod \"94d1860f-c6d0-4d65-b7b7-c70b945d9211\" (UID: \"94d1860f-c6d0-4d65-b7b7-c70b945d9211\") " Oct 03 13:54:19 crc kubenswrapper[4578]: I1003 13:54:19.682856 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlj2w\" (UniqueName: \"kubernetes.io/projected/94d1860f-c6d0-4d65-b7b7-c70b945d9211-kube-api-access-jlj2w\") pod \"94d1860f-c6d0-4d65-b7b7-c70b945d9211\" (UID: \"94d1860f-c6d0-4d65-b7b7-c70b945d9211\") " Oct 03 13:54:19 crc kubenswrapper[4578]: I1003 13:54:19.683739 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94d1860f-c6d0-4d65-b7b7-c70b945d9211-utilities" (OuterVolumeSpecName: "utilities") pod "94d1860f-c6d0-4d65-b7b7-c70b945d9211" (UID: "94d1860f-c6d0-4d65-b7b7-c70b945d9211"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:54:19 crc kubenswrapper[4578]: I1003 13:54:19.688773 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94d1860f-c6d0-4d65-b7b7-c70b945d9211-kube-api-access-jlj2w" (OuterVolumeSpecName: "kube-api-access-jlj2w") pod "94d1860f-c6d0-4d65-b7b7-c70b945d9211" (UID: "94d1860f-c6d0-4d65-b7b7-c70b945d9211"). InnerVolumeSpecName "kube-api-access-jlj2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:54:19 crc kubenswrapper[4578]: I1003 13:54:19.745254 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94d1860f-c6d0-4d65-b7b7-c70b945d9211-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94d1860f-c6d0-4d65-b7b7-c70b945d9211" (UID: "94d1860f-c6d0-4d65-b7b7-c70b945d9211"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:54:19 crc kubenswrapper[4578]: I1003 13:54:19.785097 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94d1860f-c6d0-4d65-b7b7-c70b945d9211-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:19 crc kubenswrapper[4578]: I1003 13:54:19.785130 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94d1860f-c6d0-4d65-b7b7-c70b945d9211-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:19 crc kubenswrapper[4578]: I1003 13:54:19.785143 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlj2w\" (UniqueName: \"kubernetes.io/projected/94d1860f-c6d0-4d65-b7b7-c70b945d9211-kube-api-access-jlj2w\") on node \"crc\" DevicePath \"\"" Oct 03 13:54:20 crc kubenswrapper[4578]: I1003 13:54:20.052865 4578 generic.go:334] "Generic (PLEG): container finished" podID="94d1860f-c6d0-4d65-b7b7-c70b945d9211" containerID="f47c30e91474c3a53994aca525e4e26a8a07f910737215f83b86c3b04617c359" exitCode=0 Oct 03 13:54:20 crc kubenswrapper[4578]: I1003 13:54:20.052906 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sv9qf" event={"ID":"94d1860f-c6d0-4d65-b7b7-c70b945d9211","Type":"ContainerDied","Data":"f47c30e91474c3a53994aca525e4e26a8a07f910737215f83b86c3b04617c359"} Oct 03 13:54:20 crc kubenswrapper[4578]: I1003 13:54:20.052930 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sv9qf" event={"ID":"94d1860f-c6d0-4d65-b7b7-c70b945d9211","Type":"ContainerDied","Data":"de62b087c08995d678b2e9d8ba6c3e9a1bd311b6b8613843f647561bfa816872"} Oct 03 13:54:20 crc kubenswrapper[4578]: I1003 13:54:20.052946 4578 scope.go:117] "RemoveContainer" containerID="f47c30e91474c3a53994aca525e4e26a8a07f910737215f83b86c3b04617c359" Oct 03 13:54:20 crc kubenswrapper[4578]: I1003 13:54:20.053094 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sv9qf" Oct 03 13:54:20 crc kubenswrapper[4578]: I1003 13:54:20.082418 4578 scope.go:117] "RemoveContainer" containerID="6971daa88c697335ce3a470e5c72f2a7ee7f2073e99d21b81e7f0d4791f2a052" Oct 03 13:54:20 crc kubenswrapper[4578]: I1003 13:54:20.086936 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sv9qf"] Oct 03 13:54:20 crc kubenswrapper[4578]: I1003 13:54:20.099598 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sv9qf"] Oct 03 13:54:20 crc kubenswrapper[4578]: I1003 13:54:20.109592 4578 scope.go:117] "RemoveContainer" containerID="8de6621d073ab8ec23594007c4317571e435e0b9750ed78498c1a2aec8b87c51" Oct 03 13:54:20 crc kubenswrapper[4578]: I1003 13:54:20.143923 4578 scope.go:117] "RemoveContainer" containerID="f47c30e91474c3a53994aca525e4e26a8a07f910737215f83b86c3b04617c359" Oct 03 13:54:20 crc kubenswrapper[4578]: E1003 13:54:20.144409 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f47c30e91474c3a53994aca525e4e26a8a07f910737215f83b86c3b04617c359\": container with ID starting with f47c30e91474c3a53994aca525e4e26a8a07f910737215f83b86c3b04617c359 not found: ID does not exist" containerID="f47c30e91474c3a53994aca525e4e26a8a07f910737215f83b86c3b04617c359" Oct 03 13:54:20 crc kubenswrapper[4578]: I1003 13:54:20.144438 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f47c30e91474c3a53994aca525e4e26a8a07f910737215f83b86c3b04617c359"} err="failed to get container status \"f47c30e91474c3a53994aca525e4e26a8a07f910737215f83b86c3b04617c359\": rpc error: code = NotFound desc = could not find container \"f47c30e91474c3a53994aca525e4e26a8a07f910737215f83b86c3b04617c359\": container with ID starting with f47c30e91474c3a53994aca525e4e26a8a07f910737215f83b86c3b04617c359 not found: ID does not exist" Oct 03 13:54:20 crc kubenswrapper[4578]: I1003 13:54:20.144461 4578 scope.go:117] "RemoveContainer" containerID="6971daa88c697335ce3a470e5c72f2a7ee7f2073e99d21b81e7f0d4791f2a052" Oct 03 13:54:20 crc kubenswrapper[4578]: E1003 13:54:20.144857 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6971daa88c697335ce3a470e5c72f2a7ee7f2073e99d21b81e7f0d4791f2a052\": container with ID starting with 6971daa88c697335ce3a470e5c72f2a7ee7f2073e99d21b81e7f0d4791f2a052 not found: ID does not exist" containerID="6971daa88c697335ce3a470e5c72f2a7ee7f2073e99d21b81e7f0d4791f2a052" Oct 03 13:54:20 crc kubenswrapper[4578]: I1003 13:54:20.144881 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6971daa88c697335ce3a470e5c72f2a7ee7f2073e99d21b81e7f0d4791f2a052"} err="failed to get container status \"6971daa88c697335ce3a470e5c72f2a7ee7f2073e99d21b81e7f0d4791f2a052\": rpc error: code = NotFound desc = could not find container \"6971daa88c697335ce3a470e5c72f2a7ee7f2073e99d21b81e7f0d4791f2a052\": container with ID starting with 6971daa88c697335ce3a470e5c72f2a7ee7f2073e99d21b81e7f0d4791f2a052 not found: ID does not exist" Oct 03 13:54:20 crc kubenswrapper[4578]: I1003 13:54:20.144895 4578 scope.go:117] "RemoveContainer" containerID="8de6621d073ab8ec23594007c4317571e435e0b9750ed78498c1a2aec8b87c51" Oct 03 13:54:20 crc kubenswrapper[4578]: E1003 13:54:20.145175 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8de6621d073ab8ec23594007c4317571e435e0b9750ed78498c1a2aec8b87c51\": container with ID starting with 8de6621d073ab8ec23594007c4317571e435e0b9750ed78498c1a2aec8b87c51 not found: ID does not exist" containerID="8de6621d073ab8ec23594007c4317571e435e0b9750ed78498c1a2aec8b87c51" Oct 03 13:54:20 crc kubenswrapper[4578]: I1003 13:54:20.145215 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8de6621d073ab8ec23594007c4317571e435e0b9750ed78498c1a2aec8b87c51"} err="failed to get container status \"8de6621d073ab8ec23594007c4317571e435e0b9750ed78498c1a2aec8b87c51\": rpc error: code = NotFound desc = could not find container \"8de6621d073ab8ec23594007c4317571e435e0b9750ed78498c1a2aec8b87c51\": container with ID starting with 8de6621d073ab8ec23594007c4317571e435e0b9750ed78498c1a2aec8b87c51 not found: ID does not exist" Oct 03 13:54:20 crc kubenswrapper[4578]: I1003 13:54:20.918160 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94d1860f-c6d0-4d65-b7b7-c70b945d9211" path="/var/lib/kubelet/pods/94d1860f-c6d0-4d65-b7b7-c70b945d9211/volumes" Oct 03 13:54:24 crc kubenswrapper[4578]: I1003 13:54:24.914941 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:54:24 crc kubenswrapper[4578]: E1003 13:54:24.917084 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:54:37 crc kubenswrapper[4578]: I1003 13:54:37.909741 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:54:37 crc kubenswrapper[4578]: E1003 13:54:37.911283 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:54:52 crc kubenswrapper[4578]: I1003 13:54:52.909766 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:54:52 crc kubenswrapper[4578]: E1003 13:54:52.911609 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:55:04 crc kubenswrapper[4578]: I1003 13:55:04.909394 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:55:04 crc kubenswrapper[4578]: E1003 13:55:04.910292 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:55:18 crc kubenswrapper[4578]: I1003 13:55:18.909887 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:55:18 crc kubenswrapper[4578]: E1003 13:55:18.910708 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:55:25 crc kubenswrapper[4578]: I1003 13:55:25.725240 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-57dc876bc5-gvzk7" podUID="bd42e3ab-aa85-453e-abfd-7b6aa55e9674" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Oct 03 13:55:30 crc kubenswrapper[4578]: I1003 13:55:30.909001 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:55:30 crc kubenswrapper[4578]: E1003 13:55:30.909784 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 13:55:42 crc kubenswrapper[4578]: I1003 13:55:42.909291 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:55:43 crc kubenswrapper[4578]: I1003 13:55:43.770750 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"2b75b852c100095e2c125693ce2dc2b822f88b011f83622b817b7249d5a44a1c"} Oct 03 13:58:05 crc kubenswrapper[4578]: I1003 13:58:05.091307 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:58:05 crc kubenswrapper[4578]: I1003 13:58:05.092880 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:58:35 crc kubenswrapper[4578]: I1003 13:58:35.091118 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:58:35 crc kubenswrapper[4578]: I1003 13:58:35.092027 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:59:05 crc kubenswrapper[4578]: I1003 13:59:05.091265 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 13:59:05 crc kubenswrapper[4578]: I1003 13:59:05.091880 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 13:59:05 crc kubenswrapper[4578]: I1003 13:59:05.091923 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 13:59:05 crc kubenswrapper[4578]: I1003 13:59:05.092596 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2b75b852c100095e2c125693ce2dc2b822f88b011f83622b817b7249d5a44a1c"} pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 13:59:05 crc kubenswrapper[4578]: I1003 13:59:05.092663 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" containerID="cri-o://2b75b852c100095e2c125693ce2dc2b822f88b011f83622b817b7249d5a44a1c" gracePeriod=600 Oct 03 13:59:05 crc kubenswrapper[4578]: I1003 13:59:05.716992 4578 generic.go:334] "Generic (PLEG): container finished" podID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerID="2b75b852c100095e2c125693ce2dc2b822f88b011f83622b817b7249d5a44a1c" exitCode=0 Oct 03 13:59:05 crc kubenswrapper[4578]: I1003 13:59:05.717085 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerDied","Data":"2b75b852c100095e2c125693ce2dc2b822f88b011f83622b817b7249d5a44a1c"} Oct 03 13:59:05 crc kubenswrapper[4578]: I1003 13:59:05.717601 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6"} Oct 03 13:59:05 crc kubenswrapper[4578]: I1003 13:59:05.717673 4578 scope.go:117] "RemoveContainer" containerID="ab5eaff084dc6242fd846e34d3d4cc1f5f0b29132962eaea0313c1fed56ef91e" Oct 03 13:59:16 crc kubenswrapper[4578]: I1003 13:59:16.845434 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5q9v6"] Oct 03 13:59:16 crc kubenswrapper[4578]: E1003 13:59:16.846467 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94d1860f-c6d0-4d65-b7b7-c70b945d9211" containerName="extract-content" Oct 03 13:59:16 crc kubenswrapper[4578]: I1003 13:59:16.846485 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="94d1860f-c6d0-4d65-b7b7-c70b945d9211" containerName="extract-content" Oct 03 13:59:16 crc kubenswrapper[4578]: E1003 13:59:16.846517 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94d1860f-c6d0-4d65-b7b7-c70b945d9211" containerName="extract-utilities" Oct 03 13:59:16 crc kubenswrapper[4578]: I1003 13:59:16.846526 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="94d1860f-c6d0-4d65-b7b7-c70b945d9211" containerName="extract-utilities" Oct 03 13:59:16 crc kubenswrapper[4578]: E1003 13:59:16.846556 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94d1860f-c6d0-4d65-b7b7-c70b945d9211" containerName="registry-server" Oct 03 13:59:16 crc kubenswrapper[4578]: I1003 13:59:16.846565 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="94d1860f-c6d0-4d65-b7b7-c70b945d9211" containerName="registry-server" Oct 03 13:59:16 crc kubenswrapper[4578]: I1003 13:59:16.847199 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="94d1860f-c6d0-4d65-b7b7-c70b945d9211" containerName="registry-server" Oct 03 13:59:16 crc kubenswrapper[4578]: I1003 13:59:16.848962 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5q9v6" Oct 03 13:59:16 crc kubenswrapper[4578]: I1003 13:59:16.881342 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5q9v6"] Oct 03 13:59:16 crc kubenswrapper[4578]: I1003 13:59:16.936655 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64fe8eb1-c733-492e-aea9-c89c5c92a72d-catalog-content\") pod \"certified-operators-5q9v6\" (UID: \"64fe8eb1-c733-492e-aea9-c89c5c92a72d\") " pod="openshift-marketplace/certified-operators-5q9v6" Oct 03 13:59:16 crc kubenswrapper[4578]: I1003 13:59:16.936720 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f89f2\" (UniqueName: \"kubernetes.io/projected/64fe8eb1-c733-492e-aea9-c89c5c92a72d-kube-api-access-f89f2\") pod \"certified-operators-5q9v6\" (UID: \"64fe8eb1-c733-492e-aea9-c89c5c92a72d\") " pod="openshift-marketplace/certified-operators-5q9v6" Oct 03 13:59:16 crc kubenswrapper[4578]: I1003 13:59:16.936790 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64fe8eb1-c733-492e-aea9-c89c5c92a72d-utilities\") pod \"certified-operators-5q9v6\" (UID: \"64fe8eb1-c733-492e-aea9-c89c5c92a72d\") " pod="openshift-marketplace/certified-operators-5q9v6" Oct 03 13:59:17 crc kubenswrapper[4578]: I1003 13:59:17.038098 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64fe8eb1-c733-492e-aea9-c89c5c92a72d-utilities\") pod \"certified-operators-5q9v6\" (UID: \"64fe8eb1-c733-492e-aea9-c89c5c92a72d\") " pod="openshift-marketplace/certified-operators-5q9v6" Oct 03 13:59:17 crc kubenswrapper[4578]: I1003 13:59:17.039069 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64fe8eb1-c733-492e-aea9-c89c5c92a72d-utilities\") pod \"certified-operators-5q9v6\" (UID: \"64fe8eb1-c733-492e-aea9-c89c5c92a72d\") " pod="openshift-marketplace/certified-operators-5q9v6" Oct 03 13:59:17 crc kubenswrapper[4578]: I1003 13:59:17.039503 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64fe8eb1-c733-492e-aea9-c89c5c92a72d-catalog-content\") pod \"certified-operators-5q9v6\" (UID: \"64fe8eb1-c733-492e-aea9-c89c5c92a72d\") " pod="openshift-marketplace/certified-operators-5q9v6" Oct 03 13:59:17 crc kubenswrapper[4578]: I1003 13:59:17.039965 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f89f2\" (UniqueName: \"kubernetes.io/projected/64fe8eb1-c733-492e-aea9-c89c5c92a72d-kube-api-access-f89f2\") pod \"certified-operators-5q9v6\" (UID: \"64fe8eb1-c733-492e-aea9-c89c5c92a72d\") " pod="openshift-marketplace/certified-operators-5q9v6" Oct 03 13:59:17 crc kubenswrapper[4578]: I1003 13:59:17.040120 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64fe8eb1-c733-492e-aea9-c89c5c92a72d-catalog-content\") pod \"certified-operators-5q9v6\" (UID: \"64fe8eb1-c733-492e-aea9-c89c5c92a72d\") " pod="openshift-marketplace/certified-operators-5q9v6" Oct 03 13:59:17 crc kubenswrapper[4578]: I1003 13:59:17.061523 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f89f2\" (UniqueName: \"kubernetes.io/projected/64fe8eb1-c733-492e-aea9-c89c5c92a72d-kube-api-access-f89f2\") pod \"certified-operators-5q9v6\" (UID: \"64fe8eb1-c733-492e-aea9-c89c5c92a72d\") " pod="openshift-marketplace/certified-operators-5q9v6" Oct 03 13:59:17 crc kubenswrapper[4578]: I1003 13:59:17.192676 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5q9v6" Oct 03 13:59:17 crc kubenswrapper[4578]: I1003 13:59:17.732100 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5q9v6"] Oct 03 13:59:17 crc kubenswrapper[4578]: W1003 13:59:17.746771 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64fe8eb1_c733_492e_aea9_c89c5c92a72d.slice/crio-68e3c9384d035d4389bd4e905ffde469ba16085bf38f45ce18310b48ffbb4d12 WatchSource:0}: Error finding container 68e3c9384d035d4389bd4e905ffde469ba16085bf38f45ce18310b48ffbb4d12: Status 404 returned error can't find the container with id 68e3c9384d035d4389bd4e905ffde469ba16085bf38f45ce18310b48ffbb4d12 Oct 03 13:59:17 crc kubenswrapper[4578]: I1003 13:59:17.829293 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5q9v6" event={"ID":"64fe8eb1-c733-492e-aea9-c89c5c92a72d","Type":"ContainerStarted","Data":"68e3c9384d035d4389bd4e905ffde469ba16085bf38f45ce18310b48ffbb4d12"} Oct 03 13:59:18 crc kubenswrapper[4578]: I1003 13:59:18.840914 4578 generic.go:334] "Generic (PLEG): container finished" podID="64fe8eb1-c733-492e-aea9-c89c5c92a72d" containerID="9c711b836b4a6b67efb18e5ea67aae3de98ce4de68993c1705699351962a8042" exitCode=0 Oct 03 13:59:18 crc kubenswrapper[4578]: I1003 13:59:18.841138 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5q9v6" event={"ID":"64fe8eb1-c733-492e-aea9-c89c5c92a72d","Type":"ContainerDied","Data":"9c711b836b4a6b67efb18e5ea67aae3de98ce4de68993c1705699351962a8042"} Oct 03 13:59:18 crc kubenswrapper[4578]: I1003 13:59:18.844457 4578 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 13:59:20 crc kubenswrapper[4578]: I1003 13:59:20.861579 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5q9v6" event={"ID":"64fe8eb1-c733-492e-aea9-c89c5c92a72d","Type":"ContainerStarted","Data":"0e02f5a203a19389162675b89f2e293d9ac0071a9c65a8db49a6b4c4017eafb1"} Oct 03 13:59:21 crc kubenswrapper[4578]: I1003 13:59:21.871196 4578 generic.go:334] "Generic (PLEG): container finished" podID="64fe8eb1-c733-492e-aea9-c89c5c92a72d" containerID="0e02f5a203a19389162675b89f2e293d9ac0071a9c65a8db49a6b4c4017eafb1" exitCode=0 Oct 03 13:59:21 crc kubenswrapper[4578]: I1003 13:59:21.871294 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5q9v6" event={"ID":"64fe8eb1-c733-492e-aea9-c89c5c92a72d","Type":"ContainerDied","Data":"0e02f5a203a19389162675b89f2e293d9ac0071a9c65a8db49a6b4c4017eafb1"} Oct 03 13:59:22 crc kubenswrapper[4578]: I1003 13:59:22.882036 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5q9v6" event={"ID":"64fe8eb1-c733-492e-aea9-c89c5c92a72d","Type":"ContainerStarted","Data":"ec24a83cd046d5ca9e79e699fc0680178cdc19de481f37dbcaf6f466d106bd4b"} Oct 03 13:59:22 crc kubenswrapper[4578]: I1003 13:59:22.902550 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5q9v6" podStartSLOduration=3.31142605 podStartE2EDuration="6.902530097s" podCreationTimestamp="2025-10-03 13:59:16 +0000 UTC" firstStartedPulling="2025-10-03 13:59:18.84409884 +0000 UTC m=+4094.642571034" lastFinishedPulling="2025-10-03 13:59:22.435202897 +0000 UTC m=+4098.233675081" observedRunningTime="2025-10-03 13:59:22.898103489 +0000 UTC m=+4098.696575673" watchObservedRunningTime="2025-10-03 13:59:22.902530097 +0000 UTC m=+4098.701002281" Oct 03 13:59:27 crc kubenswrapper[4578]: I1003 13:59:27.194527 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5q9v6" Oct 03 13:59:27 crc kubenswrapper[4578]: I1003 13:59:27.195027 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5q9v6" Oct 03 13:59:27 crc kubenswrapper[4578]: I1003 13:59:27.238851 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5q9v6" Oct 03 13:59:27 crc kubenswrapper[4578]: I1003 13:59:27.971312 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5q9v6" Oct 03 13:59:28 crc kubenswrapper[4578]: I1003 13:59:28.026877 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5q9v6"] Oct 03 13:59:29 crc kubenswrapper[4578]: I1003 13:59:29.942510 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5q9v6" podUID="64fe8eb1-c733-492e-aea9-c89c5c92a72d" containerName="registry-server" containerID="cri-o://ec24a83cd046d5ca9e79e699fc0680178cdc19de481f37dbcaf6f466d106bd4b" gracePeriod=2 Oct 03 13:59:30 crc kubenswrapper[4578]: I1003 13:59:30.375146 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5q9v6" Oct 03 13:59:30 crc kubenswrapper[4578]: I1003 13:59:30.491573 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f89f2\" (UniqueName: \"kubernetes.io/projected/64fe8eb1-c733-492e-aea9-c89c5c92a72d-kube-api-access-f89f2\") pod \"64fe8eb1-c733-492e-aea9-c89c5c92a72d\" (UID: \"64fe8eb1-c733-492e-aea9-c89c5c92a72d\") " Oct 03 13:59:30 crc kubenswrapper[4578]: I1003 13:59:30.491694 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64fe8eb1-c733-492e-aea9-c89c5c92a72d-utilities\") pod \"64fe8eb1-c733-492e-aea9-c89c5c92a72d\" (UID: \"64fe8eb1-c733-492e-aea9-c89c5c92a72d\") " Oct 03 13:59:30 crc kubenswrapper[4578]: I1003 13:59:30.491816 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64fe8eb1-c733-492e-aea9-c89c5c92a72d-catalog-content\") pod \"64fe8eb1-c733-492e-aea9-c89c5c92a72d\" (UID: \"64fe8eb1-c733-492e-aea9-c89c5c92a72d\") " Oct 03 13:59:30 crc kubenswrapper[4578]: I1003 13:59:30.492731 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64fe8eb1-c733-492e-aea9-c89c5c92a72d-utilities" (OuterVolumeSpecName: "utilities") pod "64fe8eb1-c733-492e-aea9-c89c5c92a72d" (UID: "64fe8eb1-c733-492e-aea9-c89c5c92a72d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:59:30 crc kubenswrapper[4578]: I1003 13:59:30.499015 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64fe8eb1-c733-492e-aea9-c89c5c92a72d-kube-api-access-f89f2" (OuterVolumeSpecName: "kube-api-access-f89f2") pod "64fe8eb1-c733-492e-aea9-c89c5c92a72d" (UID: "64fe8eb1-c733-492e-aea9-c89c5c92a72d"). InnerVolumeSpecName "kube-api-access-f89f2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 13:59:30 crc kubenswrapper[4578]: I1003 13:59:30.537421 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64fe8eb1-c733-492e-aea9-c89c5c92a72d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "64fe8eb1-c733-492e-aea9-c89c5c92a72d" (UID: "64fe8eb1-c733-492e-aea9-c89c5c92a72d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 13:59:30 crc kubenswrapper[4578]: I1003 13:59:30.594525 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64fe8eb1-c733-492e-aea9-c89c5c92a72d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 13:59:30 crc kubenswrapper[4578]: I1003 13:59:30.594567 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f89f2\" (UniqueName: \"kubernetes.io/projected/64fe8eb1-c733-492e-aea9-c89c5c92a72d-kube-api-access-f89f2\") on node \"crc\" DevicePath \"\"" Oct 03 13:59:30 crc kubenswrapper[4578]: I1003 13:59:30.594581 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64fe8eb1-c733-492e-aea9-c89c5c92a72d-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 13:59:30 crc kubenswrapper[4578]: I1003 13:59:30.957006 4578 generic.go:334] "Generic (PLEG): container finished" podID="64fe8eb1-c733-492e-aea9-c89c5c92a72d" containerID="ec24a83cd046d5ca9e79e699fc0680178cdc19de481f37dbcaf6f466d106bd4b" exitCode=0 Oct 03 13:59:30 crc kubenswrapper[4578]: I1003 13:59:30.957058 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5q9v6" event={"ID":"64fe8eb1-c733-492e-aea9-c89c5c92a72d","Type":"ContainerDied","Data":"ec24a83cd046d5ca9e79e699fc0680178cdc19de481f37dbcaf6f466d106bd4b"} Oct 03 13:59:30 crc kubenswrapper[4578]: I1003 13:59:30.957089 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5q9v6" event={"ID":"64fe8eb1-c733-492e-aea9-c89c5c92a72d","Type":"ContainerDied","Data":"68e3c9384d035d4389bd4e905ffde469ba16085bf38f45ce18310b48ffbb4d12"} Oct 03 13:59:30 crc kubenswrapper[4578]: I1003 13:59:30.957111 4578 scope.go:117] "RemoveContainer" containerID="ec24a83cd046d5ca9e79e699fc0680178cdc19de481f37dbcaf6f466d106bd4b" Oct 03 13:59:30 crc kubenswrapper[4578]: I1003 13:59:30.957260 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5q9v6" Oct 03 13:59:30 crc kubenswrapper[4578]: I1003 13:59:30.997499 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5q9v6"] Oct 03 13:59:31 crc kubenswrapper[4578]: I1003 13:59:31.003345 4578 scope.go:117] "RemoveContainer" containerID="0e02f5a203a19389162675b89f2e293d9ac0071a9c65a8db49a6b4c4017eafb1" Oct 03 13:59:31 crc kubenswrapper[4578]: I1003 13:59:31.006748 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5q9v6"] Oct 03 13:59:31 crc kubenswrapper[4578]: I1003 13:59:31.029831 4578 scope.go:117] "RemoveContainer" containerID="9c711b836b4a6b67efb18e5ea67aae3de98ce4de68993c1705699351962a8042" Oct 03 13:59:31 crc kubenswrapper[4578]: I1003 13:59:31.075728 4578 scope.go:117] "RemoveContainer" containerID="ec24a83cd046d5ca9e79e699fc0680178cdc19de481f37dbcaf6f466d106bd4b" Oct 03 13:59:31 crc kubenswrapper[4578]: E1003 13:59:31.076395 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec24a83cd046d5ca9e79e699fc0680178cdc19de481f37dbcaf6f466d106bd4b\": container with ID starting with ec24a83cd046d5ca9e79e699fc0680178cdc19de481f37dbcaf6f466d106bd4b not found: ID does not exist" containerID="ec24a83cd046d5ca9e79e699fc0680178cdc19de481f37dbcaf6f466d106bd4b" Oct 03 13:59:31 crc kubenswrapper[4578]: I1003 13:59:31.076451 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec24a83cd046d5ca9e79e699fc0680178cdc19de481f37dbcaf6f466d106bd4b"} err="failed to get container status \"ec24a83cd046d5ca9e79e699fc0680178cdc19de481f37dbcaf6f466d106bd4b\": rpc error: code = NotFound desc = could not find container \"ec24a83cd046d5ca9e79e699fc0680178cdc19de481f37dbcaf6f466d106bd4b\": container with ID starting with ec24a83cd046d5ca9e79e699fc0680178cdc19de481f37dbcaf6f466d106bd4b not found: ID does not exist" Oct 03 13:59:31 crc kubenswrapper[4578]: I1003 13:59:31.076477 4578 scope.go:117] "RemoveContainer" containerID="0e02f5a203a19389162675b89f2e293d9ac0071a9c65a8db49a6b4c4017eafb1" Oct 03 13:59:31 crc kubenswrapper[4578]: E1003 13:59:31.076939 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e02f5a203a19389162675b89f2e293d9ac0071a9c65a8db49a6b4c4017eafb1\": container with ID starting with 0e02f5a203a19389162675b89f2e293d9ac0071a9c65a8db49a6b4c4017eafb1 not found: ID does not exist" containerID="0e02f5a203a19389162675b89f2e293d9ac0071a9c65a8db49a6b4c4017eafb1" Oct 03 13:59:31 crc kubenswrapper[4578]: I1003 13:59:31.076983 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e02f5a203a19389162675b89f2e293d9ac0071a9c65a8db49a6b4c4017eafb1"} err="failed to get container status \"0e02f5a203a19389162675b89f2e293d9ac0071a9c65a8db49a6b4c4017eafb1\": rpc error: code = NotFound desc = could not find container \"0e02f5a203a19389162675b89f2e293d9ac0071a9c65a8db49a6b4c4017eafb1\": container with ID starting with 0e02f5a203a19389162675b89f2e293d9ac0071a9c65a8db49a6b4c4017eafb1 not found: ID does not exist" Oct 03 13:59:31 crc kubenswrapper[4578]: I1003 13:59:31.077008 4578 scope.go:117] "RemoveContainer" containerID="9c711b836b4a6b67efb18e5ea67aae3de98ce4de68993c1705699351962a8042" Oct 03 13:59:31 crc kubenswrapper[4578]: E1003 13:59:31.077331 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c711b836b4a6b67efb18e5ea67aae3de98ce4de68993c1705699351962a8042\": container with ID starting with 9c711b836b4a6b67efb18e5ea67aae3de98ce4de68993c1705699351962a8042 not found: ID does not exist" containerID="9c711b836b4a6b67efb18e5ea67aae3de98ce4de68993c1705699351962a8042" Oct 03 13:59:31 crc kubenswrapper[4578]: I1003 13:59:31.077362 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c711b836b4a6b67efb18e5ea67aae3de98ce4de68993c1705699351962a8042"} err="failed to get container status \"9c711b836b4a6b67efb18e5ea67aae3de98ce4de68993c1705699351962a8042\": rpc error: code = NotFound desc = could not find container \"9c711b836b4a6b67efb18e5ea67aae3de98ce4de68993c1705699351962a8042\": container with ID starting with 9c711b836b4a6b67efb18e5ea67aae3de98ce4de68993c1705699351962a8042 not found: ID does not exist" Oct 03 13:59:32 crc kubenswrapper[4578]: I1003 13:59:32.918667 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64fe8eb1-c733-492e-aea9-c89c5c92a72d" path="/var/lib/kubelet/pods/64fe8eb1-c733-492e-aea9-c89c5c92a72d/volumes" Oct 03 14:00:00 crc kubenswrapper[4578]: I1003 14:00:00.152794 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8"] Oct 03 14:00:00 crc kubenswrapper[4578]: E1003 14:00:00.155141 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64fe8eb1-c733-492e-aea9-c89c5c92a72d" containerName="registry-server" Oct 03 14:00:00 crc kubenswrapper[4578]: I1003 14:00:00.155295 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="64fe8eb1-c733-492e-aea9-c89c5c92a72d" containerName="registry-server" Oct 03 14:00:00 crc kubenswrapper[4578]: E1003 14:00:00.155425 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64fe8eb1-c733-492e-aea9-c89c5c92a72d" containerName="extract-content" Oct 03 14:00:00 crc kubenswrapper[4578]: I1003 14:00:00.155561 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="64fe8eb1-c733-492e-aea9-c89c5c92a72d" containerName="extract-content" Oct 03 14:00:00 crc kubenswrapper[4578]: E1003 14:00:00.155761 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64fe8eb1-c733-492e-aea9-c89c5c92a72d" containerName="extract-utilities" Oct 03 14:00:00 crc kubenswrapper[4578]: I1003 14:00:00.155922 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="64fe8eb1-c733-492e-aea9-c89c5c92a72d" containerName="extract-utilities" Oct 03 14:00:00 crc kubenswrapper[4578]: I1003 14:00:00.156391 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="64fe8eb1-c733-492e-aea9-c89c5c92a72d" containerName="registry-server" Oct 03 14:00:00 crc kubenswrapper[4578]: I1003 14:00:00.157504 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8" Oct 03 14:00:00 crc kubenswrapper[4578]: I1003 14:00:00.160777 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 14:00:00 crc kubenswrapper[4578]: I1003 14:00:00.160832 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 14:00:00 crc kubenswrapper[4578]: I1003 14:00:00.178220 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8"] Oct 03 14:00:00 crc kubenswrapper[4578]: I1003 14:00:00.277737 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffdzq\" (UniqueName: \"kubernetes.io/projected/7948857d-7008-41b4-8107-b8d550c9c0a0-kube-api-access-ffdzq\") pod \"collect-profiles-29325000-ltbp8\" (UID: \"7948857d-7008-41b4-8107-b8d550c9c0a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8" Oct 03 14:00:00 crc kubenswrapper[4578]: I1003 14:00:00.277784 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7948857d-7008-41b4-8107-b8d550c9c0a0-config-volume\") pod \"collect-profiles-29325000-ltbp8\" (UID: \"7948857d-7008-41b4-8107-b8d550c9c0a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8" Oct 03 14:00:00 crc kubenswrapper[4578]: I1003 14:00:00.277851 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7948857d-7008-41b4-8107-b8d550c9c0a0-secret-volume\") pod \"collect-profiles-29325000-ltbp8\" (UID: \"7948857d-7008-41b4-8107-b8d550c9c0a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8" Oct 03 14:00:00 crc kubenswrapper[4578]: I1003 14:00:00.379757 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7948857d-7008-41b4-8107-b8d550c9c0a0-secret-volume\") pod \"collect-profiles-29325000-ltbp8\" (UID: \"7948857d-7008-41b4-8107-b8d550c9c0a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8" Oct 03 14:00:00 crc kubenswrapper[4578]: I1003 14:00:00.379961 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffdzq\" (UniqueName: \"kubernetes.io/projected/7948857d-7008-41b4-8107-b8d550c9c0a0-kube-api-access-ffdzq\") pod \"collect-profiles-29325000-ltbp8\" (UID: \"7948857d-7008-41b4-8107-b8d550c9c0a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8" Oct 03 14:00:00 crc kubenswrapper[4578]: I1003 14:00:00.379982 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7948857d-7008-41b4-8107-b8d550c9c0a0-config-volume\") pod \"collect-profiles-29325000-ltbp8\" (UID: \"7948857d-7008-41b4-8107-b8d550c9c0a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8" Oct 03 14:00:00 crc kubenswrapper[4578]: I1003 14:00:00.380811 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7948857d-7008-41b4-8107-b8d550c9c0a0-config-volume\") pod \"collect-profiles-29325000-ltbp8\" (UID: \"7948857d-7008-41b4-8107-b8d550c9c0a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8" Oct 03 14:00:00 crc kubenswrapper[4578]: I1003 14:00:00.385791 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7948857d-7008-41b4-8107-b8d550c9c0a0-secret-volume\") pod \"collect-profiles-29325000-ltbp8\" (UID: \"7948857d-7008-41b4-8107-b8d550c9c0a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8" Oct 03 14:00:00 crc kubenswrapper[4578]: I1003 14:00:00.397419 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffdzq\" (UniqueName: \"kubernetes.io/projected/7948857d-7008-41b4-8107-b8d550c9c0a0-kube-api-access-ffdzq\") pod \"collect-profiles-29325000-ltbp8\" (UID: \"7948857d-7008-41b4-8107-b8d550c9c0a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8" Oct 03 14:00:00 crc kubenswrapper[4578]: I1003 14:00:00.483251 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8" Oct 03 14:00:00 crc kubenswrapper[4578]: I1003 14:00:00.938052 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8"] Oct 03 14:00:01 crc kubenswrapper[4578]: I1003 14:00:01.222492 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8" event={"ID":"7948857d-7008-41b4-8107-b8d550c9c0a0","Type":"ContainerStarted","Data":"398561a927813f39bab9b9ac8a5d4ab561b0bfaa836ad25fa181574a5ab86a91"} Oct 03 14:00:01 crc kubenswrapper[4578]: I1003 14:00:01.223938 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8" event={"ID":"7948857d-7008-41b4-8107-b8d550c9c0a0","Type":"ContainerStarted","Data":"1be8269633befa85e9d98fcb1d311dbae13b37fbbe7195a4942e8e7ccf43eaa8"} Oct 03 14:00:01 crc kubenswrapper[4578]: I1003 14:00:01.238319 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8" podStartSLOduration=1.238297055 podStartE2EDuration="1.238297055s" podCreationTimestamp="2025-10-03 14:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:00:01.237186981 +0000 UTC m=+4137.035659185" watchObservedRunningTime="2025-10-03 14:00:01.238297055 +0000 UTC m=+4137.036769249" Oct 03 14:00:02 crc kubenswrapper[4578]: I1003 14:00:02.232651 4578 generic.go:334] "Generic (PLEG): container finished" podID="7948857d-7008-41b4-8107-b8d550c9c0a0" containerID="398561a927813f39bab9b9ac8a5d4ab561b0bfaa836ad25fa181574a5ab86a91" exitCode=0 Oct 03 14:00:02 crc kubenswrapper[4578]: I1003 14:00:02.232704 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8" event={"ID":"7948857d-7008-41b4-8107-b8d550c9c0a0","Type":"ContainerDied","Data":"398561a927813f39bab9b9ac8a5d4ab561b0bfaa836ad25fa181574a5ab86a91"} Oct 03 14:00:04 crc kubenswrapper[4578]: I1003 14:00:04.023420 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8" Oct 03 14:00:04 crc kubenswrapper[4578]: I1003 14:00:04.177040 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7948857d-7008-41b4-8107-b8d550c9c0a0-secret-volume\") pod \"7948857d-7008-41b4-8107-b8d550c9c0a0\" (UID: \"7948857d-7008-41b4-8107-b8d550c9c0a0\") " Oct 03 14:00:04 crc kubenswrapper[4578]: I1003 14:00:04.177541 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7948857d-7008-41b4-8107-b8d550c9c0a0-config-volume\") pod \"7948857d-7008-41b4-8107-b8d550c9c0a0\" (UID: \"7948857d-7008-41b4-8107-b8d550c9c0a0\") " Oct 03 14:00:04 crc kubenswrapper[4578]: I1003 14:00:04.178381 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7948857d-7008-41b4-8107-b8d550c9c0a0-config-volume" (OuterVolumeSpecName: "config-volume") pod "7948857d-7008-41b4-8107-b8d550c9c0a0" (UID: "7948857d-7008-41b4-8107-b8d550c9c0a0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 14:00:04 crc kubenswrapper[4578]: I1003 14:00:04.178503 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffdzq\" (UniqueName: \"kubernetes.io/projected/7948857d-7008-41b4-8107-b8d550c9c0a0-kube-api-access-ffdzq\") pod \"7948857d-7008-41b4-8107-b8d550c9c0a0\" (UID: \"7948857d-7008-41b4-8107-b8d550c9c0a0\") " Oct 03 14:00:04 crc kubenswrapper[4578]: I1003 14:00:04.179740 4578 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7948857d-7008-41b4-8107-b8d550c9c0a0-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 14:00:04 crc kubenswrapper[4578]: I1003 14:00:04.199285 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7948857d-7008-41b4-8107-b8d550c9c0a0-kube-api-access-ffdzq" (OuterVolumeSpecName: "kube-api-access-ffdzq") pod "7948857d-7008-41b4-8107-b8d550c9c0a0" (UID: "7948857d-7008-41b4-8107-b8d550c9c0a0"). InnerVolumeSpecName "kube-api-access-ffdzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:00:04 crc kubenswrapper[4578]: I1003 14:00:04.199727 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7948857d-7008-41b4-8107-b8d550c9c0a0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7948857d-7008-41b4-8107-b8d550c9c0a0" (UID: "7948857d-7008-41b4-8107-b8d550c9c0a0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 14:00:04 crc kubenswrapper[4578]: I1003 14:00:04.249431 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8" event={"ID":"7948857d-7008-41b4-8107-b8d550c9c0a0","Type":"ContainerDied","Data":"1be8269633befa85e9d98fcb1d311dbae13b37fbbe7195a4942e8e7ccf43eaa8"} Oct 03 14:00:04 crc kubenswrapper[4578]: I1003 14:00:04.249507 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1be8269633befa85e9d98fcb1d311dbae13b37fbbe7195a4942e8e7ccf43eaa8" Oct 03 14:00:04 crc kubenswrapper[4578]: I1003 14:00:04.249583 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325000-ltbp8" Oct 03 14:00:04 crc kubenswrapper[4578]: I1003 14:00:04.280819 4578 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7948857d-7008-41b4-8107-b8d550c9c0a0-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 14:00:04 crc kubenswrapper[4578]: I1003 14:00:04.280846 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffdzq\" (UniqueName: \"kubernetes.io/projected/7948857d-7008-41b4-8107-b8d550c9c0a0-kube-api-access-ffdzq\") on node \"crc\" DevicePath \"\"" Oct 03 14:00:05 crc kubenswrapper[4578]: I1003 14:00:05.104397 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz"] Oct 03 14:00:05 crc kubenswrapper[4578]: I1003 14:00:05.112322 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324955-rk7qz"] Oct 03 14:00:06 crc kubenswrapper[4578]: I1003 14:00:06.920185 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f002fd1d-bac7-406b-ba17-bd3d68d2052b" path="/var/lib/kubelet/pods/f002fd1d-bac7-406b-ba17-bd3d68d2052b/volumes" Oct 03 14:00:10 crc kubenswrapper[4578]: I1003 14:00:10.229396 4578 scope.go:117] "RemoveContainer" containerID="3629c9811602657c37fee60ede1c360c942f31c728f4b833eaebb5494327cb9a" Oct 03 14:00:32 crc kubenswrapper[4578]: I1003 14:00:32.736273 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nrl48"] Oct 03 14:00:32 crc kubenswrapper[4578]: E1003 14:00:32.737152 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7948857d-7008-41b4-8107-b8d550c9c0a0" containerName="collect-profiles" Oct 03 14:00:32 crc kubenswrapper[4578]: I1003 14:00:32.737165 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="7948857d-7008-41b4-8107-b8d550c9c0a0" containerName="collect-profiles" Oct 03 14:00:32 crc kubenswrapper[4578]: I1003 14:00:32.737356 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="7948857d-7008-41b4-8107-b8d550c9c0a0" containerName="collect-profiles" Oct 03 14:00:32 crc kubenswrapper[4578]: I1003 14:00:32.738584 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nrl48" Oct 03 14:00:32 crc kubenswrapper[4578]: I1003 14:00:32.754172 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nrl48"] Oct 03 14:00:32 crc kubenswrapper[4578]: I1003 14:00:32.845944 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5hph\" (UniqueName: \"kubernetes.io/projected/86b46c8a-86ee-46a3-8a0c-c09b6a176bbf-kube-api-access-d5hph\") pod \"redhat-operators-nrl48\" (UID: \"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf\") " pod="openshift-marketplace/redhat-operators-nrl48" Oct 03 14:00:32 crc kubenswrapper[4578]: I1003 14:00:32.845999 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86b46c8a-86ee-46a3-8a0c-c09b6a176bbf-catalog-content\") pod \"redhat-operators-nrl48\" (UID: \"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf\") " pod="openshift-marketplace/redhat-operators-nrl48" Oct 03 14:00:32 crc kubenswrapper[4578]: I1003 14:00:32.846042 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86b46c8a-86ee-46a3-8a0c-c09b6a176bbf-utilities\") pod \"redhat-operators-nrl48\" (UID: \"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf\") " pod="openshift-marketplace/redhat-operators-nrl48" Oct 03 14:00:32 crc kubenswrapper[4578]: I1003 14:00:32.948376 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86b46c8a-86ee-46a3-8a0c-c09b6a176bbf-utilities\") pod \"redhat-operators-nrl48\" (UID: \"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf\") " pod="openshift-marketplace/redhat-operators-nrl48" Oct 03 14:00:32 crc kubenswrapper[4578]: I1003 14:00:32.948560 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5hph\" (UniqueName: \"kubernetes.io/projected/86b46c8a-86ee-46a3-8a0c-c09b6a176bbf-kube-api-access-d5hph\") pod \"redhat-operators-nrl48\" (UID: \"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf\") " pod="openshift-marketplace/redhat-operators-nrl48" Oct 03 14:00:32 crc kubenswrapper[4578]: I1003 14:00:32.948594 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86b46c8a-86ee-46a3-8a0c-c09b6a176bbf-catalog-content\") pod \"redhat-operators-nrl48\" (UID: \"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf\") " pod="openshift-marketplace/redhat-operators-nrl48" Oct 03 14:00:32 crc kubenswrapper[4578]: I1003 14:00:32.949103 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86b46c8a-86ee-46a3-8a0c-c09b6a176bbf-catalog-content\") pod \"redhat-operators-nrl48\" (UID: \"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf\") " pod="openshift-marketplace/redhat-operators-nrl48" Oct 03 14:00:32 crc kubenswrapper[4578]: I1003 14:00:32.949105 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86b46c8a-86ee-46a3-8a0c-c09b6a176bbf-utilities\") pod \"redhat-operators-nrl48\" (UID: \"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf\") " pod="openshift-marketplace/redhat-operators-nrl48" Oct 03 14:00:32 crc kubenswrapper[4578]: I1003 14:00:32.976960 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5hph\" (UniqueName: \"kubernetes.io/projected/86b46c8a-86ee-46a3-8a0c-c09b6a176bbf-kube-api-access-d5hph\") pod \"redhat-operators-nrl48\" (UID: \"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf\") " pod="openshift-marketplace/redhat-operators-nrl48" Oct 03 14:00:33 crc kubenswrapper[4578]: I1003 14:00:33.120271 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nrl48" Oct 03 14:00:33 crc kubenswrapper[4578]: I1003 14:00:33.701783 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nrl48"] Oct 03 14:00:34 crc kubenswrapper[4578]: I1003 14:00:34.564775 4578 generic.go:334] "Generic (PLEG): container finished" podID="86b46c8a-86ee-46a3-8a0c-c09b6a176bbf" containerID="bdbee48110af860714f837c83ba1aae0c3e071229163bd12fd3576bb2ae0f3bf" exitCode=0 Oct 03 14:00:34 crc kubenswrapper[4578]: I1003 14:00:34.564847 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrl48" event={"ID":"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf","Type":"ContainerDied","Data":"bdbee48110af860714f837c83ba1aae0c3e071229163bd12fd3576bb2ae0f3bf"} Oct 03 14:00:34 crc kubenswrapper[4578]: I1003 14:00:34.565102 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrl48" event={"ID":"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf","Type":"ContainerStarted","Data":"939203322e7ab4fb900f52b63b13e77847d70d8e513932da2790f2805987d058"} Oct 03 14:00:36 crc kubenswrapper[4578]: I1003 14:00:36.604051 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrl48" event={"ID":"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf","Type":"ContainerStarted","Data":"f4515034dba995d831085de0f531c29f51c3c0dbccfa43bbd9edc5774ecdbf16"} Oct 03 14:00:40 crc kubenswrapper[4578]: E1003 14:00:40.965082 4578 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86b46c8a_86ee_46a3_8a0c_c09b6a176bbf.slice/crio-conmon-f4515034dba995d831085de0f531c29f51c3c0dbccfa43bbd9edc5774ecdbf16.scope\": RecentStats: unable to find data in memory cache]" Oct 03 14:00:41 crc kubenswrapper[4578]: I1003 14:00:41.665805 4578 generic.go:334] "Generic (PLEG): container finished" podID="86b46c8a-86ee-46a3-8a0c-c09b6a176bbf" containerID="f4515034dba995d831085de0f531c29f51c3c0dbccfa43bbd9edc5774ecdbf16" exitCode=0 Oct 03 14:00:41 crc kubenswrapper[4578]: I1003 14:00:41.665886 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrl48" event={"ID":"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf","Type":"ContainerDied","Data":"f4515034dba995d831085de0f531c29f51c3c0dbccfa43bbd9edc5774ecdbf16"} Oct 03 14:00:42 crc kubenswrapper[4578]: I1003 14:00:42.681569 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrl48" event={"ID":"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf","Type":"ContainerStarted","Data":"47108cc9e473110ce259eb44fb241c947835d1828b44e185eb38bb042e83a0f9"} Oct 03 14:00:42 crc kubenswrapper[4578]: I1003 14:00:42.698766 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nrl48" podStartSLOduration=3.149703422 podStartE2EDuration="10.698751946s" podCreationTimestamp="2025-10-03 14:00:32 +0000 UTC" firstStartedPulling="2025-10-03 14:00:34.566706778 +0000 UTC m=+4170.365178962" lastFinishedPulling="2025-10-03 14:00:42.115755302 +0000 UTC m=+4177.914227486" observedRunningTime="2025-10-03 14:00:42.695514515 +0000 UTC m=+4178.493986709" watchObservedRunningTime="2025-10-03 14:00:42.698751946 +0000 UTC m=+4178.497224130" Oct 03 14:00:43 crc kubenswrapper[4578]: I1003 14:00:43.120957 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nrl48" Oct 03 14:00:43 crc kubenswrapper[4578]: I1003 14:00:43.121457 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nrl48" Oct 03 14:00:44 crc kubenswrapper[4578]: I1003 14:00:44.178334 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nrl48" podUID="86b46c8a-86ee-46a3-8a0c-c09b6a176bbf" containerName="registry-server" probeResult="failure" output=< Oct 03 14:00:44 crc kubenswrapper[4578]: timeout: failed to connect service ":50051" within 1s Oct 03 14:00:44 crc kubenswrapper[4578]: > Oct 03 14:00:53 crc kubenswrapper[4578]: I1003 14:00:53.167365 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nrl48" Oct 03 14:00:53 crc kubenswrapper[4578]: I1003 14:00:53.219013 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nrl48" Oct 03 14:00:54 crc kubenswrapper[4578]: I1003 14:00:54.098541 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nrl48"] Oct 03 14:00:54 crc kubenswrapper[4578]: I1003 14:00:54.777882 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nrl48" podUID="86b46c8a-86ee-46a3-8a0c-c09b6a176bbf" containerName="registry-server" containerID="cri-o://47108cc9e473110ce259eb44fb241c947835d1828b44e185eb38bb042e83a0f9" gracePeriod=2 Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.202198 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nrl48" Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.310596 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86b46c8a-86ee-46a3-8a0c-c09b6a176bbf-catalog-content\") pod \"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf\" (UID: \"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf\") " Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.310895 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5hph\" (UniqueName: \"kubernetes.io/projected/86b46c8a-86ee-46a3-8a0c-c09b6a176bbf-kube-api-access-d5hph\") pod \"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf\" (UID: \"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf\") " Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.310985 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86b46c8a-86ee-46a3-8a0c-c09b6a176bbf-utilities\") pod \"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf\" (UID: \"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf\") " Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.311748 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86b46c8a-86ee-46a3-8a0c-c09b6a176bbf-utilities" (OuterVolumeSpecName: "utilities") pod "86b46c8a-86ee-46a3-8a0c-c09b6a176bbf" (UID: "86b46c8a-86ee-46a3-8a0c-c09b6a176bbf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.318847 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86b46c8a-86ee-46a3-8a0c-c09b6a176bbf-kube-api-access-d5hph" (OuterVolumeSpecName: "kube-api-access-d5hph") pod "86b46c8a-86ee-46a3-8a0c-c09b6a176bbf" (UID: "86b46c8a-86ee-46a3-8a0c-c09b6a176bbf"). InnerVolumeSpecName "kube-api-access-d5hph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.405052 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86b46c8a-86ee-46a3-8a0c-c09b6a176bbf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "86b46c8a-86ee-46a3-8a0c-c09b6a176bbf" (UID: "86b46c8a-86ee-46a3-8a0c-c09b6a176bbf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.412757 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86b46c8a-86ee-46a3-8a0c-c09b6a176bbf-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.412794 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86b46c8a-86ee-46a3-8a0c-c09b6a176bbf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.412810 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5hph\" (UniqueName: \"kubernetes.io/projected/86b46c8a-86ee-46a3-8a0c-c09b6a176bbf-kube-api-access-d5hph\") on node \"crc\" DevicePath \"\"" Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.786424 4578 generic.go:334] "Generic (PLEG): container finished" podID="86b46c8a-86ee-46a3-8a0c-c09b6a176bbf" containerID="47108cc9e473110ce259eb44fb241c947835d1828b44e185eb38bb042e83a0f9" exitCode=0 Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.786466 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nrl48" Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.786478 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrl48" event={"ID":"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf","Type":"ContainerDied","Data":"47108cc9e473110ce259eb44fb241c947835d1828b44e185eb38bb042e83a0f9"} Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.786509 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrl48" event={"ID":"86b46c8a-86ee-46a3-8a0c-c09b6a176bbf","Type":"ContainerDied","Data":"939203322e7ab4fb900f52b63b13e77847d70d8e513932da2790f2805987d058"} Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.786531 4578 scope.go:117] "RemoveContainer" containerID="47108cc9e473110ce259eb44fb241c947835d1828b44e185eb38bb042e83a0f9" Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.808726 4578 scope.go:117] "RemoveContainer" containerID="f4515034dba995d831085de0f531c29f51c3c0dbccfa43bbd9edc5774ecdbf16" Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.830129 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nrl48"] Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.834999 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nrl48"] Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.837234 4578 scope.go:117] "RemoveContainer" containerID="bdbee48110af860714f837c83ba1aae0c3e071229163bd12fd3576bb2ae0f3bf" Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.914313 4578 scope.go:117] "RemoveContainer" containerID="47108cc9e473110ce259eb44fb241c947835d1828b44e185eb38bb042e83a0f9" Oct 03 14:00:55 crc kubenswrapper[4578]: E1003 14:00:55.914930 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47108cc9e473110ce259eb44fb241c947835d1828b44e185eb38bb042e83a0f9\": container with ID starting with 47108cc9e473110ce259eb44fb241c947835d1828b44e185eb38bb042e83a0f9 not found: ID does not exist" containerID="47108cc9e473110ce259eb44fb241c947835d1828b44e185eb38bb042e83a0f9" Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.915055 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47108cc9e473110ce259eb44fb241c947835d1828b44e185eb38bb042e83a0f9"} err="failed to get container status \"47108cc9e473110ce259eb44fb241c947835d1828b44e185eb38bb042e83a0f9\": rpc error: code = NotFound desc = could not find container \"47108cc9e473110ce259eb44fb241c947835d1828b44e185eb38bb042e83a0f9\": container with ID starting with 47108cc9e473110ce259eb44fb241c947835d1828b44e185eb38bb042e83a0f9 not found: ID does not exist" Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.915085 4578 scope.go:117] "RemoveContainer" containerID="f4515034dba995d831085de0f531c29f51c3c0dbccfa43bbd9edc5774ecdbf16" Oct 03 14:00:55 crc kubenswrapper[4578]: E1003 14:00:55.916242 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4515034dba995d831085de0f531c29f51c3c0dbccfa43bbd9edc5774ecdbf16\": container with ID starting with f4515034dba995d831085de0f531c29f51c3c0dbccfa43bbd9edc5774ecdbf16 not found: ID does not exist" containerID="f4515034dba995d831085de0f531c29f51c3c0dbccfa43bbd9edc5774ecdbf16" Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.916280 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4515034dba995d831085de0f531c29f51c3c0dbccfa43bbd9edc5774ecdbf16"} err="failed to get container status \"f4515034dba995d831085de0f531c29f51c3c0dbccfa43bbd9edc5774ecdbf16\": rpc error: code = NotFound desc = could not find container \"f4515034dba995d831085de0f531c29f51c3c0dbccfa43bbd9edc5774ecdbf16\": container with ID starting with f4515034dba995d831085de0f531c29f51c3c0dbccfa43bbd9edc5774ecdbf16 not found: ID does not exist" Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.916301 4578 scope.go:117] "RemoveContainer" containerID="bdbee48110af860714f837c83ba1aae0c3e071229163bd12fd3576bb2ae0f3bf" Oct 03 14:00:55 crc kubenswrapper[4578]: E1003 14:00:55.916600 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdbee48110af860714f837c83ba1aae0c3e071229163bd12fd3576bb2ae0f3bf\": container with ID starting with bdbee48110af860714f837c83ba1aae0c3e071229163bd12fd3576bb2ae0f3bf not found: ID does not exist" containerID="bdbee48110af860714f837c83ba1aae0c3e071229163bd12fd3576bb2ae0f3bf" Oct 03 14:00:55 crc kubenswrapper[4578]: I1003 14:00:55.916651 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdbee48110af860714f837c83ba1aae0c3e071229163bd12fd3576bb2ae0f3bf"} err="failed to get container status \"bdbee48110af860714f837c83ba1aae0c3e071229163bd12fd3576bb2ae0f3bf\": rpc error: code = NotFound desc = could not find container \"bdbee48110af860714f837c83ba1aae0c3e071229163bd12fd3576bb2ae0f3bf\": container with ID starting with bdbee48110af860714f837c83ba1aae0c3e071229163bd12fd3576bb2ae0f3bf not found: ID does not exist" Oct 03 14:00:56 crc kubenswrapper[4578]: I1003 14:00:56.922033 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86b46c8a-86ee-46a3-8a0c-c09b6a176bbf" path="/var/lib/kubelet/pods/86b46c8a-86ee-46a3-8a0c-c09b6a176bbf/volumes" Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.150405 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29325001-hhn29"] Oct 03 14:01:00 crc kubenswrapper[4578]: E1003 14:01:00.151338 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86b46c8a-86ee-46a3-8a0c-c09b6a176bbf" containerName="extract-utilities" Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.151352 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="86b46c8a-86ee-46a3-8a0c-c09b6a176bbf" containerName="extract-utilities" Oct 03 14:01:00 crc kubenswrapper[4578]: E1003 14:01:00.151367 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86b46c8a-86ee-46a3-8a0c-c09b6a176bbf" containerName="registry-server" Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.151374 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="86b46c8a-86ee-46a3-8a0c-c09b6a176bbf" containerName="registry-server" Oct 03 14:01:00 crc kubenswrapper[4578]: E1003 14:01:00.151394 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86b46c8a-86ee-46a3-8a0c-c09b6a176bbf" containerName="extract-content" Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.151402 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="86b46c8a-86ee-46a3-8a0c-c09b6a176bbf" containerName="extract-content" Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.151589 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="86b46c8a-86ee-46a3-8a0c-c09b6a176bbf" containerName="registry-server" Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.152221 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325001-hhn29" Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.172860 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29325001-hhn29"] Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.302990 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-fernet-keys\") pod \"keystone-cron-29325001-hhn29\" (UID: \"7d6d4828-38c5-41cf-93a4-a25d8a9810a2\") " pod="openstack/keystone-cron-29325001-hhn29" Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.303079 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-config-data\") pod \"keystone-cron-29325001-hhn29\" (UID: \"7d6d4828-38c5-41cf-93a4-a25d8a9810a2\") " pod="openstack/keystone-cron-29325001-hhn29" Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.303150 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-combined-ca-bundle\") pod \"keystone-cron-29325001-hhn29\" (UID: \"7d6d4828-38c5-41cf-93a4-a25d8a9810a2\") " pod="openstack/keystone-cron-29325001-hhn29" Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.303221 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxgvt\" (UniqueName: \"kubernetes.io/projected/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-kube-api-access-vxgvt\") pod \"keystone-cron-29325001-hhn29\" (UID: \"7d6d4828-38c5-41cf-93a4-a25d8a9810a2\") " pod="openstack/keystone-cron-29325001-hhn29" Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.405280 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxgvt\" (UniqueName: \"kubernetes.io/projected/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-kube-api-access-vxgvt\") pod \"keystone-cron-29325001-hhn29\" (UID: \"7d6d4828-38c5-41cf-93a4-a25d8a9810a2\") " pod="openstack/keystone-cron-29325001-hhn29" Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.405399 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-fernet-keys\") pod \"keystone-cron-29325001-hhn29\" (UID: \"7d6d4828-38c5-41cf-93a4-a25d8a9810a2\") " pod="openstack/keystone-cron-29325001-hhn29" Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.405462 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-config-data\") pod \"keystone-cron-29325001-hhn29\" (UID: \"7d6d4828-38c5-41cf-93a4-a25d8a9810a2\") " pod="openstack/keystone-cron-29325001-hhn29" Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.405533 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-combined-ca-bundle\") pod \"keystone-cron-29325001-hhn29\" (UID: \"7d6d4828-38c5-41cf-93a4-a25d8a9810a2\") " pod="openstack/keystone-cron-29325001-hhn29" Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.411330 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-fernet-keys\") pod \"keystone-cron-29325001-hhn29\" (UID: \"7d6d4828-38c5-41cf-93a4-a25d8a9810a2\") " pod="openstack/keystone-cron-29325001-hhn29" Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.412518 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-combined-ca-bundle\") pod \"keystone-cron-29325001-hhn29\" (UID: \"7d6d4828-38c5-41cf-93a4-a25d8a9810a2\") " pod="openstack/keystone-cron-29325001-hhn29" Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.413244 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-config-data\") pod \"keystone-cron-29325001-hhn29\" (UID: \"7d6d4828-38c5-41cf-93a4-a25d8a9810a2\") " pod="openstack/keystone-cron-29325001-hhn29" Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.424852 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxgvt\" (UniqueName: \"kubernetes.io/projected/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-kube-api-access-vxgvt\") pod \"keystone-cron-29325001-hhn29\" (UID: \"7d6d4828-38c5-41cf-93a4-a25d8a9810a2\") " pod="openstack/keystone-cron-29325001-hhn29" Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.484122 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325001-hhn29" Oct 03 14:01:00 crc kubenswrapper[4578]: I1003 14:01:00.951363 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29325001-hhn29"] Oct 03 14:01:01 crc kubenswrapper[4578]: I1003 14:01:01.849622 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325001-hhn29" event={"ID":"7d6d4828-38c5-41cf-93a4-a25d8a9810a2","Type":"ContainerStarted","Data":"65468e0e23adc603f3b3d5bc366e1892d5aef4bf42eac77df6046a3ae87eeab4"} Oct 03 14:01:01 crc kubenswrapper[4578]: I1003 14:01:01.850014 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325001-hhn29" event={"ID":"7d6d4828-38c5-41cf-93a4-a25d8a9810a2","Type":"ContainerStarted","Data":"506fb949ef6ed5022cdbf3d9de87fec37b9342e95628530c2e73a8e2a9eaee3f"} Oct 03 14:01:01 crc kubenswrapper[4578]: I1003 14:01:01.870727 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29325001-hhn29" podStartSLOduration=1.870711253 podStartE2EDuration="1.870711253s" podCreationTimestamp="2025-10-03 14:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:01:01.866493412 +0000 UTC m=+4197.664965596" watchObservedRunningTime="2025-10-03 14:01:01.870711253 +0000 UTC m=+4197.669183437" Oct 03 14:01:04 crc kubenswrapper[4578]: I1003 14:01:04.877013 4578 generic.go:334] "Generic (PLEG): container finished" podID="7d6d4828-38c5-41cf-93a4-a25d8a9810a2" containerID="65468e0e23adc603f3b3d5bc366e1892d5aef4bf42eac77df6046a3ae87eeab4" exitCode=0 Oct 03 14:01:04 crc kubenswrapper[4578]: I1003 14:01:04.877134 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325001-hhn29" event={"ID":"7d6d4828-38c5-41cf-93a4-a25d8a9810a2","Type":"ContainerDied","Data":"65468e0e23adc603f3b3d5bc366e1892d5aef4bf42eac77df6046a3ae87eeab4"} Oct 03 14:01:05 crc kubenswrapper[4578]: I1003 14:01:05.091986 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:01:05 crc kubenswrapper[4578]: I1003 14:01:05.092038 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:01:06 crc kubenswrapper[4578]: I1003 14:01:06.261713 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325001-hhn29" Oct 03 14:01:06 crc kubenswrapper[4578]: I1003 14:01:06.417024 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxgvt\" (UniqueName: \"kubernetes.io/projected/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-kube-api-access-vxgvt\") pod \"7d6d4828-38c5-41cf-93a4-a25d8a9810a2\" (UID: \"7d6d4828-38c5-41cf-93a4-a25d8a9810a2\") " Oct 03 14:01:06 crc kubenswrapper[4578]: I1003 14:01:06.417079 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-combined-ca-bundle\") pod \"7d6d4828-38c5-41cf-93a4-a25d8a9810a2\" (UID: \"7d6d4828-38c5-41cf-93a4-a25d8a9810a2\") " Oct 03 14:01:06 crc kubenswrapper[4578]: I1003 14:01:06.417206 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-fernet-keys\") pod \"7d6d4828-38c5-41cf-93a4-a25d8a9810a2\" (UID: \"7d6d4828-38c5-41cf-93a4-a25d8a9810a2\") " Oct 03 14:01:06 crc kubenswrapper[4578]: I1003 14:01:06.417268 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-config-data\") pod \"7d6d4828-38c5-41cf-93a4-a25d8a9810a2\" (UID: \"7d6d4828-38c5-41cf-93a4-a25d8a9810a2\") " Oct 03 14:01:06 crc kubenswrapper[4578]: I1003 14:01:06.423448 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-kube-api-access-vxgvt" (OuterVolumeSpecName: "kube-api-access-vxgvt") pod "7d6d4828-38c5-41cf-93a4-a25d8a9810a2" (UID: "7d6d4828-38c5-41cf-93a4-a25d8a9810a2"). InnerVolumeSpecName "kube-api-access-vxgvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:01:06 crc kubenswrapper[4578]: I1003 14:01:06.424093 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7d6d4828-38c5-41cf-93a4-a25d8a9810a2" (UID: "7d6d4828-38c5-41cf-93a4-a25d8a9810a2"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 14:01:06 crc kubenswrapper[4578]: I1003 14:01:06.456253 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d6d4828-38c5-41cf-93a4-a25d8a9810a2" (UID: "7d6d4828-38c5-41cf-93a4-a25d8a9810a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 14:01:06 crc kubenswrapper[4578]: I1003 14:01:06.476110 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-config-data" (OuterVolumeSpecName: "config-data") pod "7d6d4828-38c5-41cf-93a4-a25d8a9810a2" (UID: "7d6d4828-38c5-41cf-93a4-a25d8a9810a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 14:01:06 crc kubenswrapper[4578]: I1003 14:01:06.519435 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxgvt\" (UniqueName: \"kubernetes.io/projected/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-kube-api-access-vxgvt\") on node \"crc\" DevicePath \"\"" Oct 03 14:01:06 crc kubenswrapper[4578]: I1003 14:01:06.519484 4578 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 03 14:01:06 crc kubenswrapper[4578]: I1003 14:01:06.519498 4578 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 03 14:01:06 crc kubenswrapper[4578]: I1003 14:01:06.519510 4578 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d6d4828-38c5-41cf-93a4-a25d8a9810a2-config-data\") on node \"crc\" DevicePath \"\"" Oct 03 14:01:06 crc kubenswrapper[4578]: I1003 14:01:06.917421 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325001-hhn29" Oct 03 14:01:06 crc kubenswrapper[4578]: I1003 14:01:06.927653 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325001-hhn29" event={"ID":"7d6d4828-38c5-41cf-93a4-a25d8a9810a2","Type":"ContainerDied","Data":"506fb949ef6ed5022cdbf3d9de87fec37b9342e95628530c2e73a8e2a9eaee3f"} Oct 03 14:01:06 crc kubenswrapper[4578]: I1003 14:01:06.927876 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="506fb949ef6ed5022cdbf3d9de87fec37b9342e95628530c2e73a8e2a9eaee3f" Oct 03 14:01:35 crc kubenswrapper[4578]: I1003 14:01:35.091676 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:01:35 crc kubenswrapper[4578]: I1003 14:01:35.092207 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:02:05 crc kubenswrapper[4578]: I1003 14:02:05.092048 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:02:05 crc kubenswrapper[4578]: I1003 14:02:05.092564 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:02:05 crc kubenswrapper[4578]: I1003 14:02:05.092612 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 14:02:05 crc kubenswrapper[4578]: I1003 14:02:05.093349 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6"} pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 14:02:05 crc kubenswrapper[4578]: I1003 14:02:05.093400 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" containerID="cri-o://8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" gracePeriod=600 Oct 03 14:02:05 crc kubenswrapper[4578]: E1003 14:02:05.215518 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:02:05 crc kubenswrapper[4578]: I1003 14:02:05.397709 4578 generic.go:334] "Generic (PLEG): container finished" podID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" exitCode=0 Oct 03 14:02:05 crc kubenswrapper[4578]: I1003 14:02:05.397751 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerDied","Data":"8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6"} Oct 03 14:02:05 crc kubenswrapper[4578]: I1003 14:02:05.397784 4578 scope.go:117] "RemoveContainer" containerID="2b75b852c100095e2c125693ce2dc2b822f88b011f83622b817b7249d5a44a1c" Oct 03 14:02:05 crc kubenswrapper[4578]: I1003 14:02:05.398441 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:02:05 crc kubenswrapper[4578]: E1003 14:02:05.398718 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:02:17 crc kubenswrapper[4578]: I1003 14:02:17.909516 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:02:17 crc kubenswrapper[4578]: E1003 14:02:17.910193 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:02:32 crc kubenswrapper[4578]: I1003 14:02:32.909366 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:02:32 crc kubenswrapper[4578]: E1003 14:02:32.910259 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:02:44 crc kubenswrapper[4578]: I1003 14:02:44.916020 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:02:44 crc kubenswrapper[4578]: E1003 14:02:44.916811 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:02:57 crc kubenswrapper[4578]: I1003 14:02:57.909487 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:02:57 crc kubenswrapper[4578]: E1003 14:02:57.910231 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:03:10 crc kubenswrapper[4578]: I1003 14:03:10.909134 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:03:10 crc kubenswrapper[4578]: E1003 14:03:10.909952 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:03:25 crc kubenswrapper[4578]: I1003 14:03:25.909735 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:03:25 crc kubenswrapper[4578]: E1003 14:03:25.910566 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:03:39 crc kubenswrapper[4578]: I1003 14:03:39.909443 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:03:39 crc kubenswrapper[4578]: E1003 14:03:39.910263 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:03:51 crc kubenswrapper[4578]: I1003 14:03:51.909607 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:03:51 crc kubenswrapper[4578]: E1003 14:03:51.910324 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:04:05 crc kubenswrapper[4578]: I1003 14:04:05.908896 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:04:05 crc kubenswrapper[4578]: E1003 14:04:05.911253 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:04:17 crc kubenswrapper[4578]: I1003 14:04:17.908947 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:04:17 crc kubenswrapper[4578]: E1003 14:04:17.909700 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:04:29 crc kubenswrapper[4578]: I1003 14:04:29.908978 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:04:29 crc kubenswrapper[4578]: E1003 14:04:29.909741 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:04:40 crc kubenswrapper[4578]: I1003 14:04:40.910105 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:04:40 crc kubenswrapper[4578]: E1003 14:04:40.910982 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:04:43 crc kubenswrapper[4578]: I1003 14:04:43.593200 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nz49l"] Oct 03 14:04:43 crc kubenswrapper[4578]: E1003 14:04:43.593955 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d6d4828-38c5-41cf-93a4-a25d8a9810a2" containerName="keystone-cron" Oct 03 14:04:43 crc kubenswrapper[4578]: I1003 14:04:43.593970 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d6d4828-38c5-41cf-93a4-a25d8a9810a2" containerName="keystone-cron" Oct 03 14:04:43 crc kubenswrapper[4578]: I1003 14:04:43.594192 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d6d4828-38c5-41cf-93a4-a25d8a9810a2" containerName="keystone-cron" Oct 03 14:04:43 crc kubenswrapper[4578]: I1003 14:04:43.598698 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nz49l" Oct 03 14:04:43 crc kubenswrapper[4578]: I1003 14:04:43.632703 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nz49l"] Oct 03 14:04:43 crc kubenswrapper[4578]: I1003 14:04:43.660588 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp2hr\" (UniqueName: \"kubernetes.io/projected/e5156347-8e12-45b9-884d-213db32fcb22-kube-api-access-hp2hr\") pod \"community-operators-nz49l\" (UID: \"e5156347-8e12-45b9-884d-213db32fcb22\") " pod="openshift-marketplace/community-operators-nz49l" Oct 03 14:04:43 crc kubenswrapper[4578]: I1003 14:04:43.660725 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5156347-8e12-45b9-884d-213db32fcb22-catalog-content\") pod \"community-operators-nz49l\" (UID: \"e5156347-8e12-45b9-884d-213db32fcb22\") " pod="openshift-marketplace/community-operators-nz49l" Oct 03 14:04:43 crc kubenswrapper[4578]: I1003 14:04:43.660959 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5156347-8e12-45b9-884d-213db32fcb22-utilities\") pod \"community-operators-nz49l\" (UID: \"e5156347-8e12-45b9-884d-213db32fcb22\") " pod="openshift-marketplace/community-operators-nz49l" Oct 03 14:04:43 crc kubenswrapper[4578]: I1003 14:04:43.764621 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5156347-8e12-45b9-884d-213db32fcb22-utilities\") pod \"community-operators-nz49l\" (UID: \"e5156347-8e12-45b9-884d-213db32fcb22\") " pod="openshift-marketplace/community-operators-nz49l" Oct 03 14:04:43 crc kubenswrapper[4578]: I1003 14:04:43.764871 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp2hr\" (UniqueName: \"kubernetes.io/projected/e5156347-8e12-45b9-884d-213db32fcb22-kube-api-access-hp2hr\") pod \"community-operators-nz49l\" (UID: \"e5156347-8e12-45b9-884d-213db32fcb22\") " pod="openshift-marketplace/community-operators-nz49l" Oct 03 14:04:43 crc kubenswrapper[4578]: I1003 14:04:43.764916 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5156347-8e12-45b9-884d-213db32fcb22-catalog-content\") pod \"community-operators-nz49l\" (UID: \"e5156347-8e12-45b9-884d-213db32fcb22\") " pod="openshift-marketplace/community-operators-nz49l" Oct 03 14:04:43 crc kubenswrapper[4578]: I1003 14:04:43.765266 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5156347-8e12-45b9-884d-213db32fcb22-utilities\") pod \"community-operators-nz49l\" (UID: \"e5156347-8e12-45b9-884d-213db32fcb22\") " pod="openshift-marketplace/community-operators-nz49l" Oct 03 14:04:43 crc kubenswrapper[4578]: I1003 14:04:43.765365 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5156347-8e12-45b9-884d-213db32fcb22-catalog-content\") pod \"community-operators-nz49l\" (UID: \"e5156347-8e12-45b9-884d-213db32fcb22\") " pod="openshift-marketplace/community-operators-nz49l" Oct 03 14:04:43 crc kubenswrapper[4578]: I1003 14:04:43.791159 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp2hr\" (UniqueName: \"kubernetes.io/projected/e5156347-8e12-45b9-884d-213db32fcb22-kube-api-access-hp2hr\") pod \"community-operators-nz49l\" (UID: \"e5156347-8e12-45b9-884d-213db32fcb22\") " pod="openshift-marketplace/community-operators-nz49l" Oct 03 14:04:43 crc kubenswrapper[4578]: I1003 14:04:43.942808 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nz49l" Oct 03 14:04:44 crc kubenswrapper[4578]: I1003 14:04:44.960992 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nz49l"] Oct 03 14:04:45 crc kubenswrapper[4578]: I1003 14:04:45.911712 4578 generic.go:334] "Generic (PLEG): container finished" podID="e5156347-8e12-45b9-884d-213db32fcb22" containerID="2f3204e651c4cdd7455fcbeb6a550488f4af8759e4c62c1f070b65156dfa7dbb" exitCode=0 Oct 03 14:04:45 crc kubenswrapper[4578]: I1003 14:04:45.911800 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nz49l" event={"ID":"e5156347-8e12-45b9-884d-213db32fcb22","Type":"ContainerDied","Data":"2f3204e651c4cdd7455fcbeb6a550488f4af8759e4c62c1f070b65156dfa7dbb"} Oct 03 14:04:45 crc kubenswrapper[4578]: I1003 14:04:45.912429 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nz49l" event={"ID":"e5156347-8e12-45b9-884d-213db32fcb22","Type":"ContainerStarted","Data":"3006e7456d2a5cc8e7a90710e490f408a47ee841e8a7c2d836947d8257bbaa2a"} Oct 03 14:04:45 crc kubenswrapper[4578]: I1003 14:04:45.916309 4578 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 14:04:47 crc kubenswrapper[4578]: I1003 14:04:47.934372 4578 generic.go:334] "Generic (PLEG): container finished" podID="e5156347-8e12-45b9-884d-213db32fcb22" containerID="90ad76dacce7f09ef0bb71d95f34804b89471ec6b473a4c2e977ff9ed38819a0" exitCode=0 Oct 03 14:04:47 crc kubenswrapper[4578]: I1003 14:04:47.934455 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nz49l" event={"ID":"e5156347-8e12-45b9-884d-213db32fcb22","Type":"ContainerDied","Data":"90ad76dacce7f09ef0bb71d95f34804b89471ec6b473a4c2e977ff9ed38819a0"} Oct 03 14:04:48 crc kubenswrapper[4578]: I1003 14:04:48.948494 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nz49l" event={"ID":"e5156347-8e12-45b9-884d-213db32fcb22","Type":"ContainerStarted","Data":"ae520f00c275247ff685630f80bb2106a90e14ae719dc94ed77a935e85cf66e6"} Oct 03 14:04:48 crc kubenswrapper[4578]: I1003 14:04:48.980841 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nz49l" podStartSLOduration=3.500520895 podStartE2EDuration="5.980819013s" podCreationTimestamp="2025-10-03 14:04:43 +0000 UTC" firstStartedPulling="2025-10-03 14:04:45.915308601 +0000 UTC m=+4421.713780805" lastFinishedPulling="2025-10-03 14:04:48.395606739 +0000 UTC m=+4424.194078923" observedRunningTime="2025-10-03 14:04:48.972941547 +0000 UTC m=+4424.771413731" watchObservedRunningTime="2025-10-03 14:04:48.980819013 +0000 UTC m=+4424.779291197" Oct 03 14:04:53 crc kubenswrapper[4578]: I1003 14:04:53.943984 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nz49l" Oct 03 14:04:53 crc kubenswrapper[4578]: I1003 14:04:53.944663 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nz49l" Oct 03 14:04:54 crc kubenswrapper[4578]: I1003 14:04:54.015345 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nz49l" Oct 03 14:04:54 crc kubenswrapper[4578]: I1003 14:04:54.090702 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nz49l" Oct 03 14:04:54 crc kubenswrapper[4578]: I1003 14:04:54.254970 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nz49l"] Oct 03 14:04:55 crc kubenswrapper[4578]: I1003 14:04:55.909502 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:04:55 crc kubenswrapper[4578]: E1003 14:04:55.909781 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:04:56 crc kubenswrapper[4578]: I1003 14:04:56.025386 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nz49l" podUID="e5156347-8e12-45b9-884d-213db32fcb22" containerName="registry-server" containerID="cri-o://ae520f00c275247ff685630f80bb2106a90e14ae719dc94ed77a935e85cf66e6" gracePeriod=2 Oct 03 14:04:56 crc kubenswrapper[4578]: I1003 14:04:56.549881 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nz49l" Oct 03 14:04:56 crc kubenswrapper[4578]: I1003 14:04:56.682973 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5156347-8e12-45b9-884d-213db32fcb22-catalog-content\") pod \"e5156347-8e12-45b9-884d-213db32fcb22\" (UID: \"e5156347-8e12-45b9-884d-213db32fcb22\") " Oct 03 14:04:56 crc kubenswrapper[4578]: I1003 14:04:56.683084 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp2hr\" (UniqueName: \"kubernetes.io/projected/e5156347-8e12-45b9-884d-213db32fcb22-kube-api-access-hp2hr\") pod \"e5156347-8e12-45b9-884d-213db32fcb22\" (UID: \"e5156347-8e12-45b9-884d-213db32fcb22\") " Oct 03 14:04:56 crc kubenswrapper[4578]: I1003 14:04:56.683114 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5156347-8e12-45b9-884d-213db32fcb22-utilities\") pod \"e5156347-8e12-45b9-884d-213db32fcb22\" (UID: \"e5156347-8e12-45b9-884d-213db32fcb22\") " Oct 03 14:04:56 crc kubenswrapper[4578]: I1003 14:04:56.684161 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5156347-8e12-45b9-884d-213db32fcb22-utilities" (OuterVolumeSpecName: "utilities") pod "e5156347-8e12-45b9-884d-213db32fcb22" (UID: "e5156347-8e12-45b9-884d-213db32fcb22"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:04:56 crc kubenswrapper[4578]: I1003 14:04:56.694489 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5156347-8e12-45b9-884d-213db32fcb22-kube-api-access-hp2hr" (OuterVolumeSpecName: "kube-api-access-hp2hr") pod "e5156347-8e12-45b9-884d-213db32fcb22" (UID: "e5156347-8e12-45b9-884d-213db32fcb22"). InnerVolumeSpecName "kube-api-access-hp2hr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:04:56 crc kubenswrapper[4578]: I1003 14:04:56.748917 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5156347-8e12-45b9-884d-213db32fcb22-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5156347-8e12-45b9-884d-213db32fcb22" (UID: "e5156347-8e12-45b9-884d-213db32fcb22"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:04:56 crc kubenswrapper[4578]: I1003 14:04:56.787511 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5156347-8e12-45b9-884d-213db32fcb22-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:04:56 crc kubenswrapper[4578]: I1003 14:04:56.787808 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp2hr\" (UniqueName: \"kubernetes.io/projected/e5156347-8e12-45b9-884d-213db32fcb22-kube-api-access-hp2hr\") on node \"crc\" DevicePath \"\"" Oct 03 14:04:56 crc kubenswrapper[4578]: I1003 14:04:56.787871 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5156347-8e12-45b9-884d-213db32fcb22-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:04:57 crc kubenswrapper[4578]: I1003 14:04:57.035382 4578 generic.go:334] "Generic (PLEG): container finished" podID="e5156347-8e12-45b9-884d-213db32fcb22" containerID="ae520f00c275247ff685630f80bb2106a90e14ae719dc94ed77a935e85cf66e6" exitCode=0 Oct 03 14:04:57 crc kubenswrapper[4578]: I1003 14:04:57.035425 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nz49l" event={"ID":"e5156347-8e12-45b9-884d-213db32fcb22","Type":"ContainerDied","Data":"ae520f00c275247ff685630f80bb2106a90e14ae719dc94ed77a935e85cf66e6"} Oct 03 14:04:57 crc kubenswrapper[4578]: I1003 14:04:57.035451 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nz49l" event={"ID":"e5156347-8e12-45b9-884d-213db32fcb22","Type":"ContainerDied","Data":"3006e7456d2a5cc8e7a90710e490f408a47ee841e8a7c2d836947d8257bbaa2a"} Oct 03 14:04:57 crc kubenswrapper[4578]: I1003 14:04:57.035467 4578 scope.go:117] "RemoveContainer" containerID="ae520f00c275247ff685630f80bb2106a90e14ae719dc94ed77a935e85cf66e6" Oct 03 14:04:57 crc kubenswrapper[4578]: I1003 14:04:57.035619 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nz49l" Oct 03 14:04:57 crc kubenswrapper[4578]: I1003 14:04:57.070103 4578 scope.go:117] "RemoveContainer" containerID="90ad76dacce7f09ef0bb71d95f34804b89471ec6b473a4c2e977ff9ed38819a0" Oct 03 14:04:57 crc kubenswrapper[4578]: I1003 14:04:57.095961 4578 scope.go:117] "RemoveContainer" containerID="2f3204e651c4cdd7455fcbeb6a550488f4af8759e4c62c1f070b65156dfa7dbb" Oct 03 14:04:57 crc kubenswrapper[4578]: I1003 14:04:57.124375 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nz49l"] Oct 03 14:04:57 crc kubenswrapper[4578]: I1003 14:04:57.140473 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nz49l"] Oct 03 14:04:57 crc kubenswrapper[4578]: I1003 14:04:57.145275 4578 scope.go:117] "RemoveContainer" containerID="ae520f00c275247ff685630f80bb2106a90e14ae719dc94ed77a935e85cf66e6" Oct 03 14:04:57 crc kubenswrapper[4578]: E1003 14:04:57.146176 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae520f00c275247ff685630f80bb2106a90e14ae719dc94ed77a935e85cf66e6\": container with ID starting with ae520f00c275247ff685630f80bb2106a90e14ae719dc94ed77a935e85cf66e6 not found: ID does not exist" containerID="ae520f00c275247ff685630f80bb2106a90e14ae719dc94ed77a935e85cf66e6" Oct 03 14:04:57 crc kubenswrapper[4578]: I1003 14:04:57.146295 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae520f00c275247ff685630f80bb2106a90e14ae719dc94ed77a935e85cf66e6"} err="failed to get container status \"ae520f00c275247ff685630f80bb2106a90e14ae719dc94ed77a935e85cf66e6\": rpc error: code = NotFound desc = could not find container \"ae520f00c275247ff685630f80bb2106a90e14ae719dc94ed77a935e85cf66e6\": container with ID starting with ae520f00c275247ff685630f80bb2106a90e14ae719dc94ed77a935e85cf66e6 not found: ID does not exist" Oct 03 14:04:57 crc kubenswrapper[4578]: I1003 14:04:57.146383 4578 scope.go:117] "RemoveContainer" containerID="90ad76dacce7f09ef0bb71d95f34804b89471ec6b473a4c2e977ff9ed38819a0" Oct 03 14:04:57 crc kubenswrapper[4578]: E1003 14:04:57.148932 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90ad76dacce7f09ef0bb71d95f34804b89471ec6b473a4c2e977ff9ed38819a0\": container with ID starting with 90ad76dacce7f09ef0bb71d95f34804b89471ec6b473a4c2e977ff9ed38819a0 not found: ID does not exist" containerID="90ad76dacce7f09ef0bb71d95f34804b89471ec6b473a4c2e977ff9ed38819a0" Oct 03 14:04:57 crc kubenswrapper[4578]: I1003 14:04:57.149037 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90ad76dacce7f09ef0bb71d95f34804b89471ec6b473a4c2e977ff9ed38819a0"} err="failed to get container status \"90ad76dacce7f09ef0bb71d95f34804b89471ec6b473a4c2e977ff9ed38819a0\": rpc error: code = NotFound desc = could not find container \"90ad76dacce7f09ef0bb71d95f34804b89471ec6b473a4c2e977ff9ed38819a0\": container with ID starting with 90ad76dacce7f09ef0bb71d95f34804b89471ec6b473a4c2e977ff9ed38819a0 not found: ID does not exist" Oct 03 14:04:57 crc kubenswrapper[4578]: I1003 14:04:57.149091 4578 scope.go:117] "RemoveContainer" containerID="2f3204e651c4cdd7455fcbeb6a550488f4af8759e4c62c1f070b65156dfa7dbb" Oct 03 14:04:57 crc kubenswrapper[4578]: E1003 14:04:57.149983 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f3204e651c4cdd7455fcbeb6a550488f4af8759e4c62c1f070b65156dfa7dbb\": container with ID starting with 2f3204e651c4cdd7455fcbeb6a550488f4af8759e4c62c1f070b65156dfa7dbb not found: ID does not exist" containerID="2f3204e651c4cdd7455fcbeb6a550488f4af8759e4c62c1f070b65156dfa7dbb" Oct 03 14:04:57 crc kubenswrapper[4578]: I1003 14:04:57.150038 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f3204e651c4cdd7455fcbeb6a550488f4af8759e4c62c1f070b65156dfa7dbb"} err="failed to get container status \"2f3204e651c4cdd7455fcbeb6a550488f4af8759e4c62c1f070b65156dfa7dbb\": rpc error: code = NotFound desc = could not find container \"2f3204e651c4cdd7455fcbeb6a550488f4af8759e4c62c1f070b65156dfa7dbb\": container with ID starting with 2f3204e651c4cdd7455fcbeb6a550488f4af8759e4c62c1f070b65156dfa7dbb not found: ID does not exist" Oct 03 14:04:58 crc kubenswrapper[4578]: I1003 14:04:58.924589 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5156347-8e12-45b9-884d-213db32fcb22" path="/var/lib/kubelet/pods/e5156347-8e12-45b9-884d-213db32fcb22/volumes" Oct 03 14:05:08 crc kubenswrapper[4578]: I1003 14:05:08.910049 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:05:08 crc kubenswrapper[4578]: E1003 14:05:08.911057 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:05:20 crc kubenswrapper[4578]: I1003 14:05:20.909911 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:05:20 crc kubenswrapper[4578]: E1003 14:05:20.912058 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:05:33 crc kubenswrapper[4578]: I1003 14:05:33.909110 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:05:33 crc kubenswrapper[4578]: E1003 14:05:33.909978 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:05:45 crc kubenswrapper[4578]: I1003 14:05:45.908998 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:05:45 crc kubenswrapper[4578]: E1003 14:05:45.909827 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:05:56 crc kubenswrapper[4578]: I1003 14:05:56.909211 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:05:56 crc kubenswrapper[4578]: E1003 14:05:56.910255 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:06:07 crc kubenswrapper[4578]: I1003 14:06:07.909679 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:06:07 crc kubenswrapper[4578]: E1003 14:06:07.910669 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:06:08 crc kubenswrapper[4578]: I1003 14:06:08.836713 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zdxss"] Oct 03 14:06:08 crc kubenswrapper[4578]: E1003 14:06:08.837686 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5156347-8e12-45b9-884d-213db32fcb22" containerName="registry-server" Oct 03 14:06:08 crc kubenswrapper[4578]: I1003 14:06:08.837861 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5156347-8e12-45b9-884d-213db32fcb22" containerName="registry-server" Oct 03 14:06:08 crc kubenswrapper[4578]: E1003 14:06:08.837890 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5156347-8e12-45b9-884d-213db32fcb22" containerName="extract-content" Oct 03 14:06:08 crc kubenswrapper[4578]: I1003 14:06:08.837898 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5156347-8e12-45b9-884d-213db32fcb22" containerName="extract-content" Oct 03 14:06:08 crc kubenswrapper[4578]: E1003 14:06:08.837920 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5156347-8e12-45b9-884d-213db32fcb22" containerName="extract-utilities" Oct 03 14:06:08 crc kubenswrapper[4578]: I1003 14:06:08.837928 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5156347-8e12-45b9-884d-213db32fcb22" containerName="extract-utilities" Oct 03 14:06:08 crc kubenswrapper[4578]: I1003 14:06:08.838101 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5156347-8e12-45b9-884d-213db32fcb22" containerName="registry-server" Oct 03 14:06:08 crc kubenswrapper[4578]: I1003 14:06:08.839466 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zdxss" Oct 03 14:06:08 crc kubenswrapper[4578]: I1003 14:06:08.858229 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zdxss"] Oct 03 14:06:09 crc kubenswrapper[4578]: I1003 14:06:09.036595 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kb7jq\" (UniqueName: \"kubernetes.io/projected/fb81241b-ce07-44f7-8779-45dce597ee28-kube-api-access-kb7jq\") pod \"redhat-marketplace-zdxss\" (UID: \"fb81241b-ce07-44f7-8779-45dce597ee28\") " pod="openshift-marketplace/redhat-marketplace-zdxss" Oct 03 14:06:09 crc kubenswrapper[4578]: I1003 14:06:09.037710 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb81241b-ce07-44f7-8779-45dce597ee28-catalog-content\") pod \"redhat-marketplace-zdxss\" (UID: \"fb81241b-ce07-44f7-8779-45dce597ee28\") " pod="openshift-marketplace/redhat-marketplace-zdxss" Oct 03 14:06:09 crc kubenswrapper[4578]: I1003 14:06:09.037846 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb81241b-ce07-44f7-8779-45dce597ee28-utilities\") pod \"redhat-marketplace-zdxss\" (UID: \"fb81241b-ce07-44f7-8779-45dce597ee28\") " pod="openshift-marketplace/redhat-marketplace-zdxss" Oct 03 14:06:09 crc kubenswrapper[4578]: I1003 14:06:09.140223 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kb7jq\" (UniqueName: \"kubernetes.io/projected/fb81241b-ce07-44f7-8779-45dce597ee28-kube-api-access-kb7jq\") pod \"redhat-marketplace-zdxss\" (UID: \"fb81241b-ce07-44f7-8779-45dce597ee28\") " pod="openshift-marketplace/redhat-marketplace-zdxss" Oct 03 14:06:09 crc kubenswrapper[4578]: I1003 14:06:09.140300 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb81241b-ce07-44f7-8779-45dce597ee28-catalog-content\") pod \"redhat-marketplace-zdxss\" (UID: \"fb81241b-ce07-44f7-8779-45dce597ee28\") " pod="openshift-marketplace/redhat-marketplace-zdxss" Oct 03 14:06:09 crc kubenswrapper[4578]: I1003 14:06:09.140325 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb81241b-ce07-44f7-8779-45dce597ee28-utilities\") pod \"redhat-marketplace-zdxss\" (UID: \"fb81241b-ce07-44f7-8779-45dce597ee28\") " pod="openshift-marketplace/redhat-marketplace-zdxss" Oct 03 14:06:09 crc kubenswrapper[4578]: I1003 14:06:09.140996 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb81241b-ce07-44f7-8779-45dce597ee28-utilities\") pod \"redhat-marketplace-zdxss\" (UID: \"fb81241b-ce07-44f7-8779-45dce597ee28\") " pod="openshift-marketplace/redhat-marketplace-zdxss" Oct 03 14:06:09 crc kubenswrapper[4578]: I1003 14:06:09.141006 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb81241b-ce07-44f7-8779-45dce597ee28-catalog-content\") pod \"redhat-marketplace-zdxss\" (UID: \"fb81241b-ce07-44f7-8779-45dce597ee28\") " pod="openshift-marketplace/redhat-marketplace-zdxss" Oct 03 14:06:09 crc kubenswrapper[4578]: I1003 14:06:09.158171 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kb7jq\" (UniqueName: \"kubernetes.io/projected/fb81241b-ce07-44f7-8779-45dce597ee28-kube-api-access-kb7jq\") pod \"redhat-marketplace-zdxss\" (UID: \"fb81241b-ce07-44f7-8779-45dce597ee28\") " pod="openshift-marketplace/redhat-marketplace-zdxss" Oct 03 14:06:09 crc kubenswrapper[4578]: I1003 14:06:09.158996 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zdxss" Oct 03 14:06:09 crc kubenswrapper[4578]: I1003 14:06:09.645875 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zdxss"] Oct 03 14:06:09 crc kubenswrapper[4578]: I1003 14:06:09.724668 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zdxss" event={"ID":"fb81241b-ce07-44f7-8779-45dce597ee28","Type":"ContainerStarted","Data":"54273247f1625ccaf8683e4d8ea02a4eb726ea0a80c0d33a6df6cf574da26065"} Oct 03 14:06:10 crc kubenswrapper[4578]: I1003 14:06:10.736981 4578 generic.go:334] "Generic (PLEG): container finished" podID="fb81241b-ce07-44f7-8779-45dce597ee28" containerID="8e1a86571ff575c11c6d4e8bee9cab4da0af96fdd9fbe406ac174bd964a1b29f" exitCode=0 Oct 03 14:06:10 crc kubenswrapper[4578]: I1003 14:06:10.737026 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zdxss" event={"ID":"fb81241b-ce07-44f7-8779-45dce597ee28","Type":"ContainerDied","Data":"8e1a86571ff575c11c6d4e8bee9cab4da0af96fdd9fbe406ac174bd964a1b29f"} Oct 03 14:06:12 crc kubenswrapper[4578]: I1003 14:06:12.759189 4578 generic.go:334] "Generic (PLEG): container finished" podID="fb81241b-ce07-44f7-8779-45dce597ee28" containerID="720bfdf4890f79201def62c2cdd1d97882e30464524bfea7da3bce6752a46c2c" exitCode=0 Oct 03 14:06:12 crc kubenswrapper[4578]: I1003 14:06:12.759732 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zdxss" event={"ID":"fb81241b-ce07-44f7-8779-45dce597ee28","Type":"ContainerDied","Data":"720bfdf4890f79201def62c2cdd1d97882e30464524bfea7da3bce6752a46c2c"} Oct 03 14:06:13 crc kubenswrapper[4578]: I1003 14:06:13.772956 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zdxss" event={"ID":"fb81241b-ce07-44f7-8779-45dce597ee28","Type":"ContainerStarted","Data":"2396f491428e33e192685f5754adbde926491c1c586ad038235b2e58423d04b6"} Oct 03 14:06:13 crc kubenswrapper[4578]: I1003 14:06:13.791371 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zdxss" podStartSLOduration=3.302866208 podStartE2EDuration="5.791352581s" podCreationTimestamp="2025-10-03 14:06:08 +0000 UTC" firstStartedPulling="2025-10-03 14:06:10.739108453 +0000 UTC m=+4506.537580637" lastFinishedPulling="2025-10-03 14:06:13.227594786 +0000 UTC m=+4509.026067010" observedRunningTime="2025-10-03 14:06:13.78936327 +0000 UTC m=+4509.587835464" watchObservedRunningTime="2025-10-03 14:06:13.791352581 +0000 UTC m=+4509.589824765" Oct 03 14:06:19 crc kubenswrapper[4578]: I1003 14:06:19.159420 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zdxss" Oct 03 14:06:19 crc kubenswrapper[4578]: I1003 14:06:19.159886 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zdxss" Oct 03 14:06:19 crc kubenswrapper[4578]: I1003 14:06:19.214376 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zdxss" Oct 03 14:06:19 crc kubenswrapper[4578]: I1003 14:06:19.887606 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zdxss" Oct 03 14:06:19 crc kubenswrapper[4578]: I1003 14:06:19.935230 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zdxss"] Oct 03 14:06:20 crc kubenswrapper[4578]: I1003 14:06:20.910213 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:06:20 crc kubenswrapper[4578]: E1003 14:06:20.910717 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:06:21 crc kubenswrapper[4578]: I1003 14:06:21.860794 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zdxss" podUID="fb81241b-ce07-44f7-8779-45dce597ee28" containerName="registry-server" containerID="cri-o://2396f491428e33e192685f5754adbde926491c1c586ad038235b2e58423d04b6" gracePeriod=2 Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.287708 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zdxss" Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.415600 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb81241b-ce07-44f7-8779-45dce597ee28-catalog-content\") pod \"fb81241b-ce07-44f7-8779-45dce597ee28\" (UID: \"fb81241b-ce07-44f7-8779-45dce597ee28\") " Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.415664 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb81241b-ce07-44f7-8779-45dce597ee28-utilities\") pod \"fb81241b-ce07-44f7-8779-45dce597ee28\" (UID: \"fb81241b-ce07-44f7-8779-45dce597ee28\") " Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.415729 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kb7jq\" (UniqueName: \"kubernetes.io/projected/fb81241b-ce07-44f7-8779-45dce597ee28-kube-api-access-kb7jq\") pod \"fb81241b-ce07-44f7-8779-45dce597ee28\" (UID: \"fb81241b-ce07-44f7-8779-45dce597ee28\") " Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.417818 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb81241b-ce07-44f7-8779-45dce597ee28-utilities" (OuterVolumeSpecName: "utilities") pod "fb81241b-ce07-44f7-8779-45dce597ee28" (UID: "fb81241b-ce07-44f7-8779-45dce597ee28"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.423878 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb81241b-ce07-44f7-8779-45dce597ee28-kube-api-access-kb7jq" (OuterVolumeSpecName: "kube-api-access-kb7jq") pod "fb81241b-ce07-44f7-8779-45dce597ee28" (UID: "fb81241b-ce07-44f7-8779-45dce597ee28"). InnerVolumeSpecName "kube-api-access-kb7jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.433023 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb81241b-ce07-44f7-8779-45dce597ee28-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fb81241b-ce07-44f7-8779-45dce597ee28" (UID: "fb81241b-ce07-44f7-8779-45dce597ee28"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.519821 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb81241b-ce07-44f7-8779-45dce597ee28-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.520015 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb81241b-ce07-44f7-8779-45dce597ee28-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.520042 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kb7jq\" (UniqueName: \"kubernetes.io/projected/fb81241b-ce07-44f7-8779-45dce597ee28-kube-api-access-kb7jq\") on node \"crc\" DevicePath \"\"" Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.871302 4578 generic.go:334] "Generic (PLEG): container finished" podID="fb81241b-ce07-44f7-8779-45dce597ee28" containerID="2396f491428e33e192685f5754adbde926491c1c586ad038235b2e58423d04b6" exitCode=0 Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.871349 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zdxss" event={"ID":"fb81241b-ce07-44f7-8779-45dce597ee28","Type":"ContainerDied","Data":"2396f491428e33e192685f5754adbde926491c1c586ad038235b2e58423d04b6"} Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.871385 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zdxss" event={"ID":"fb81241b-ce07-44f7-8779-45dce597ee28","Type":"ContainerDied","Data":"54273247f1625ccaf8683e4d8ea02a4eb726ea0a80c0d33a6df6cf574da26065"} Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.871406 4578 scope.go:117] "RemoveContainer" containerID="2396f491428e33e192685f5754adbde926491c1c586ad038235b2e58423d04b6" Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.871488 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zdxss" Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.892172 4578 scope.go:117] "RemoveContainer" containerID="720bfdf4890f79201def62c2cdd1d97882e30464524bfea7da3bce6752a46c2c" Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.933918 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zdxss"] Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.940822 4578 scope.go:117] "RemoveContainer" containerID="8e1a86571ff575c11c6d4e8bee9cab4da0af96fdd9fbe406ac174bd964a1b29f" Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.944210 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zdxss"] Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.966713 4578 scope.go:117] "RemoveContainer" containerID="2396f491428e33e192685f5754adbde926491c1c586ad038235b2e58423d04b6" Oct 03 14:06:22 crc kubenswrapper[4578]: E1003 14:06:22.967259 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2396f491428e33e192685f5754adbde926491c1c586ad038235b2e58423d04b6\": container with ID starting with 2396f491428e33e192685f5754adbde926491c1c586ad038235b2e58423d04b6 not found: ID does not exist" containerID="2396f491428e33e192685f5754adbde926491c1c586ad038235b2e58423d04b6" Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.967317 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2396f491428e33e192685f5754adbde926491c1c586ad038235b2e58423d04b6"} err="failed to get container status \"2396f491428e33e192685f5754adbde926491c1c586ad038235b2e58423d04b6\": rpc error: code = NotFound desc = could not find container \"2396f491428e33e192685f5754adbde926491c1c586ad038235b2e58423d04b6\": container with ID starting with 2396f491428e33e192685f5754adbde926491c1c586ad038235b2e58423d04b6 not found: ID does not exist" Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.967349 4578 scope.go:117] "RemoveContainer" containerID="720bfdf4890f79201def62c2cdd1d97882e30464524bfea7da3bce6752a46c2c" Oct 03 14:06:22 crc kubenswrapper[4578]: E1003 14:06:22.967800 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"720bfdf4890f79201def62c2cdd1d97882e30464524bfea7da3bce6752a46c2c\": container with ID starting with 720bfdf4890f79201def62c2cdd1d97882e30464524bfea7da3bce6752a46c2c not found: ID does not exist" containerID="720bfdf4890f79201def62c2cdd1d97882e30464524bfea7da3bce6752a46c2c" Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.967834 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"720bfdf4890f79201def62c2cdd1d97882e30464524bfea7da3bce6752a46c2c"} err="failed to get container status \"720bfdf4890f79201def62c2cdd1d97882e30464524bfea7da3bce6752a46c2c\": rpc error: code = NotFound desc = could not find container \"720bfdf4890f79201def62c2cdd1d97882e30464524bfea7da3bce6752a46c2c\": container with ID starting with 720bfdf4890f79201def62c2cdd1d97882e30464524bfea7da3bce6752a46c2c not found: ID does not exist" Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.967856 4578 scope.go:117] "RemoveContainer" containerID="8e1a86571ff575c11c6d4e8bee9cab4da0af96fdd9fbe406ac174bd964a1b29f" Oct 03 14:06:22 crc kubenswrapper[4578]: E1003 14:06:22.968216 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e1a86571ff575c11c6d4e8bee9cab4da0af96fdd9fbe406ac174bd964a1b29f\": container with ID starting with 8e1a86571ff575c11c6d4e8bee9cab4da0af96fdd9fbe406ac174bd964a1b29f not found: ID does not exist" containerID="8e1a86571ff575c11c6d4e8bee9cab4da0af96fdd9fbe406ac174bd964a1b29f" Oct 03 14:06:22 crc kubenswrapper[4578]: I1003 14:06:22.968262 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e1a86571ff575c11c6d4e8bee9cab4da0af96fdd9fbe406ac174bd964a1b29f"} err="failed to get container status \"8e1a86571ff575c11c6d4e8bee9cab4da0af96fdd9fbe406ac174bd964a1b29f\": rpc error: code = NotFound desc = could not find container \"8e1a86571ff575c11c6d4e8bee9cab4da0af96fdd9fbe406ac174bd964a1b29f\": container with ID starting with 8e1a86571ff575c11c6d4e8bee9cab4da0af96fdd9fbe406ac174bd964a1b29f not found: ID does not exist" Oct 03 14:06:24 crc kubenswrapper[4578]: I1003 14:06:24.920086 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb81241b-ce07-44f7-8779-45dce597ee28" path="/var/lib/kubelet/pods/fb81241b-ce07-44f7-8779-45dce597ee28/volumes" Oct 03 14:06:32 crc kubenswrapper[4578]: I1003 14:06:32.909777 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:06:32 crc kubenswrapper[4578]: E1003 14:06:32.910568 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:06:47 crc kubenswrapper[4578]: I1003 14:06:47.909682 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:06:47 crc kubenswrapper[4578]: E1003 14:06:47.910603 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:07:00 crc kubenswrapper[4578]: I1003 14:07:00.910736 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:07:00 crc kubenswrapper[4578]: E1003 14:07:00.911910 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:07:12 crc kubenswrapper[4578]: I1003 14:07:12.909085 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:07:13 crc kubenswrapper[4578]: I1003 14:07:13.365827 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"d75edac35d3e8d13aacc39445c65e996b89e3d449792b96e3b020bb14e4e8800"} Oct 03 14:09:35 crc kubenswrapper[4578]: I1003 14:09:35.092038 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:09:35 crc kubenswrapper[4578]: I1003 14:09:35.092916 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:10:05 crc kubenswrapper[4578]: I1003 14:10:05.091429 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:10:05 crc kubenswrapper[4578]: I1003 14:10:05.091888 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:10:20 crc kubenswrapper[4578]: I1003 14:10:20.266014 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-z7wfg"] Oct 03 14:10:20 crc kubenswrapper[4578]: E1003 14:10:20.267106 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb81241b-ce07-44f7-8779-45dce597ee28" containerName="extract-content" Oct 03 14:10:20 crc kubenswrapper[4578]: I1003 14:10:20.267122 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb81241b-ce07-44f7-8779-45dce597ee28" containerName="extract-content" Oct 03 14:10:20 crc kubenswrapper[4578]: E1003 14:10:20.267142 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb81241b-ce07-44f7-8779-45dce597ee28" containerName="registry-server" Oct 03 14:10:20 crc kubenswrapper[4578]: I1003 14:10:20.267147 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb81241b-ce07-44f7-8779-45dce597ee28" containerName="registry-server" Oct 03 14:10:20 crc kubenswrapper[4578]: E1003 14:10:20.267164 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb81241b-ce07-44f7-8779-45dce597ee28" containerName="extract-utilities" Oct 03 14:10:20 crc kubenswrapper[4578]: I1003 14:10:20.267170 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb81241b-ce07-44f7-8779-45dce597ee28" containerName="extract-utilities" Oct 03 14:10:20 crc kubenswrapper[4578]: I1003 14:10:20.267340 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb81241b-ce07-44f7-8779-45dce597ee28" containerName="registry-server" Oct 03 14:10:20 crc kubenswrapper[4578]: I1003 14:10:20.271286 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z7wfg" Oct 03 14:10:20 crc kubenswrapper[4578]: I1003 14:10:20.294088 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z7wfg"] Oct 03 14:10:20 crc kubenswrapper[4578]: I1003 14:10:20.296453 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7ab277-af83-484c-917b-5398fbb64bcd-catalog-content\") pod \"certified-operators-z7wfg\" (UID: \"aa7ab277-af83-484c-917b-5398fbb64bcd\") " pod="openshift-marketplace/certified-operators-z7wfg" Oct 03 14:10:20 crc kubenswrapper[4578]: I1003 14:10:20.296556 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7ab277-af83-484c-917b-5398fbb64bcd-utilities\") pod \"certified-operators-z7wfg\" (UID: \"aa7ab277-af83-484c-917b-5398fbb64bcd\") " pod="openshift-marketplace/certified-operators-z7wfg" Oct 03 14:10:20 crc kubenswrapper[4578]: I1003 14:10:20.296615 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxk8k\" (UniqueName: \"kubernetes.io/projected/aa7ab277-af83-484c-917b-5398fbb64bcd-kube-api-access-wxk8k\") pod \"certified-operators-z7wfg\" (UID: \"aa7ab277-af83-484c-917b-5398fbb64bcd\") " pod="openshift-marketplace/certified-operators-z7wfg" Oct 03 14:10:20 crc kubenswrapper[4578]: I1003 14:10:20.397776 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxk8k\" (UniqueName: \"kubernetes.io/projected/aa7ab277-af83-484c-917b-5398fbb64bcd-kube-api-access-wxk8k\") pod \"certified-operators-z7wfg\" (UID: \"aa7ab277-af83-484c-917b-5398fbb64bcd\") " pod="openshift-marketplace/certified-operators-z7wfg" Oct 03 14:10:20 crc kubenswrapper[4578]: I1003 14:10:20.398156 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7ab277-af83-484c-917b-5398fbb64bcd-catalog-content\") pod \"certified-operators-z7wfg\" (UID: \"aa7ab277-af83-484c-917b-5398fbb64bcd\") " pod="openshift-marketplace/certified-operators-z7wfg" Oct 03 14:10:20 crc kubenswrapper[4578]: I1003 14:10:20.398307 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7ab277-af83-484c-917b-5398fbb64bcd-utilities\") pod \"certified-operators-z7wfg\" (UID: \"aa7ab277-af83-484c-917b-5398fbb64bcd\") " pod="openshift-marketplace/certified-operators-z7wfg" Oct 03 14:10:20 crc kubenswrapper[4578]: I1003 14:10:20.398773 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7ab277-af83-484c-917b-5398fbb64bcd-utilities\") pod \"certified-operators-z7wfg\" (UID: \"aa7ab277-af83-484c-917b-5398fbb64bcd\") " pod="openshift-marketplace/certified-operators-z7wfg" Oct 03 14:10:20 crc kubenswrapper[4578]: I1003 14:10:20.398780 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7ab277-af83-484c-917b-5398fbb64bcd-catalog-content\") pod \"certified-operators-z7wfg\" (UID: \"aa7ab277-af83-484c-917b-5398fbb64bcd\") " pod="openshift-marketplace/certified-operators-z7wfg" Oct 03 14:10:20 crc kubenswrapper[4578]: I1003 14:10:20.419025 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxk8k\" (UniqueName: \"kubernetes.io/projected/aa7ab277-af83-484c-917b-5398fbb64bcd-kube-api-access-wxk8k\") pod \"certified-operators-z7wfg\" (UID: \"aa7ab277-af83-484c-917b-5398fbb64bcd\") " pod="openshift-marketplace/certified-operators-z7wfg" Oct 03 14:10:20 crc kubenswrapper[4578]: I1003 14:10:20.591872 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z7wfg" Oct 03 14:10:20 crc kubenswrapper[4578]: I1003 14:10:20.942321 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z7wfg"] Oct 03 14:10:20 crc kubenswrapper[4578]: W1003 14:10:20.951339 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa7ab277_af83_484c_917b_5398fbb64bcd.slice/crio-3c51741347aacfd1d659fc92444f0fc66a5b2d7bc41e65baa0d754ef422afbe8 WatchSource:0}: Error finding container 3c51741347aacfd1d659fc92444f0fc66a5b2d7bc41e65baa0d754ef422afbe8: Status 404 returned error can't find the container with id 3c51741347aacfd1d659fc92444f0fc66a5b2d7bc41e65baa0d754ef422afbe8 Oct 03 14:10:21 crc kubenswrapper[4578]: I1003 14:10:21.055666 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z7wfg" event={"ID":"aa7ab277-af83-484c-917b-5398fbb64bcd","Type":"ContainerStarted","Data":"3c51741347aacfd1d659fc92444f0fc66a5b2d7bc41e65baa0d754ef422afbe8"} Oct 03 14:10:22 crc kubenswrapper[4578]: I1003 14:10:22.065780 4578 generic.go:334] "Generic (PLEG): container finished" podID="aa7ab277-af83-484c-917b-5398fbb64bcd" containerID="fde50f866088e5c235bd47c4e4ae9fc3bf1028aa6b27ed0827aa63d14d181d12" exitCode=0 Oct 03 14:10:22 crc kubenswrapper[4578]: I1003 14:10:22.065870 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z7wfg" event={"ID":"aa7ab277-af83-484c-917b-5398fbb64bcd","Type":"ContainerDied","Data":"fde50f866088e5c235bd47c4e4ae9fc3bf1028aa6b27ed0827aa63d14d181d12"} Oct 03 14:10:22 crc kubenswrapper[4578]: I1003 14:10:22.068137 4578 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 14:10:24 crc kubenswrapper[4578]: I1003 14:10:24.097613 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z7wfg" event={"ID":"aa7ab277-af83-484c-917b-5398fbb64bcd","Type":"ContainerStarted","Data":"191d40d7c13db3659433f41ef668c794150fc4bc36d4f113095774871130c999"} Oct 03 14:10:26 crc kubenswrapper[4578]: I1003 14:10:26.119839 4578 generic.go:334] "Generic (PLEG): container finished" podID="aa7ab277-af83-484c-917b-5398fbb64bcd" containerID="191d40d7c13db3659433f41ef668c794150fc4bc36d4f113095774871130c999" exitCode=0 Oct 03 14:10:26 crc kubenswrapper[4578]: I1003 14:10:26.119916 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z7wfg" event={"ID":"aa7ab277-af83-484c-917b-5398fbb64bcd","Type":"ContainerDied","Data":"191d40d7c13db3659433f41ef668c794150fc4bc36d4f113095774871130c999"} Oct 03 14:10:27 crc kubenswrapper[4578]: I1003 14:10:27.131339 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z7wfg" event={"ID":"aa7ab277-af83-484c-917b-5398fbb64bcd","Type":"ContainerStarted","Data":"ab05544afde008e17bd8949b6444b499b86eab4cedc3487c997c6113f6b4d231"} Oct 03 14:10:27 crc kubenswrapper[4578]: I1003 14:10:27.161060 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-z7wfg" podStartSLOduration=2.677535887 podStartE2EDuration="7.16104576s" podCreationTimestamp="2025-10-03 14:10:20 +0000 UTC" firstStartedPulling="2025-10-03 14:10:22.067940095 +0000 UTC m=+4757.866412279" lastFinishedPulling="2025-10-03 14:10:26.551449968 +0000 UTC m=+4762.349922152" observedRunningTime="2025-10-03 14:10:27.156189439 +0000 UTC m=+4762.954661633" watchObservedRunningTime="2025-10-03 14:10:27.16104576 +0000 UTC m=+4762.959517944" Oct 03 14:10:30 crc kubenswrapper[4578]: I1003 14:10:30.592660 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-z7wfg" Oct 03 14:10:30 crc kubenswrapper[4578]: I1003 14:10:30.593296 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-z7wfg" Oct 03 14:10:30 crc kubenswrapper[4578]: I1003 14:10:30.652526 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-z7wfg" Oct 03 14:10:31 crc kubenswrapper[4578]: I1003 14:10:31.218490 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-z7wfg" Oct 03 14:10:31 crc kubenswrapper[4578]: I1003 14:10:31.273768 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z7wfg"] Oct 03 14:10:33 crc kubenswrapper[4578]: I1003 14:10:33.182246 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-z7wfg" podUID="aa7ab277-af83-484c-917b-5398fbb64bcd" containerName="registry-server" containerID="cri-o://ab05544afde008e17bd8949b6444b499b86eab4cedc3487c997c6113f6b4d231" gracePeriod=2 Oct 03 14:10:33 crc kubenswrapper[4578]: I1003 14:10:33.635988 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z7wfg" Oct 03 14:10:33 crc kubenswrapper[4578]: I1003 14:10:33.645872 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxk8k\" (UniqueName: \"kubernetes.io/projected/aa7ab277-af83-484c-917b-5398fbb64bcd-kube-api-access-wxk8k\") pod \"aa7ab277-af83-484c-917b-5398fbb64bcd\" (UID: \"aa7ab277-af83-484c-917b-5398fbb64bcd\") " Oct 03 14:10:33 crc kubenswrapper[4578]: I1003 14:10:33.645958 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7ab277-af83-484c-917b-5398fbb64bcd-utilities\") pod \"aa7ab277-af83-484c-917b-5398fbb64bcd\" (UID: \"aa7ab277-af83-484c-917b-5398fbb64bcd\") " Oct 03 14:10:33 crc kubenswrapper[4578]: I1003 14:10:33.646026 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7ab277-af83-484c-917b-5398fbb64bcd-catalog-content\") pod \"aa7ab277-af83-484c-917b-5398fbb64bcd\" (UID: \"aa7ab277-af83-484c-917b-5398fbb64bcd\") " Oct 03 14:10:33 crc kubenswrapper[4578]: I1003 14:10:33.647189 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa7ab277-af83-484c-917b-5398fbb64bcd-utilities" (OuterVolumeSpecName: "utilities") pod "aa7ab277-af83-484c-917b-5398fbb64bcd" (UID: "aa7ab277-af83-484c-917b-5398fbb64bcd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:10:33 crc kubenswrapper[4578]: I1003 14:10:33.655900 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa7ab277-af83-484c-917b-5398fbb64bcd-kube-api-access-wxk8k" (OuterVolumeSpecName: "kube-api-access-wxk8k") pod "aa7ab277-af83-484c-917b-5398fbb64bcd" (UID: "aa7ab277-af83-484c-917b-5398fbb64bcd"). InnerVolumeSpecName "kube-api-access-wxk8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:10:33 crc kubenswrapper[4578]: I1003 14:10:33.712707 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa7ab277-af83-484c-917b-5398fbb64bcd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa7ab277-af83-484c-917b-5398fbb64bcd" (UID: "aa7ab277-af83-484c-917b-5398fbb64bcd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:10:33 crc kubenswrapper[4578]: I1003 14:10:33.747891 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxk8k\" (UniqueName: \"kubernetes.io/projected/aa7ab277-af83-484c-917b-5398fbb64bcd-kube-api-access-wxk8k\") on node \"crc\" DevicePath \"\"" Oct 03 14:10:33 crc kubenswrapper[4578]: I1003 14:10:33.747921 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7ab277-af83-484c-917b-5398fbb64bcd-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:10:33 crc kubenswrapper[4578]: I1003 14:10:33.747933 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7ab277-af83-484c-917b-5398fbb64bcd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:10:34 crc kubenswrapper[4578]: I1003 14:10:34.196291 4578 generic.go:334] "Generic (PLEG): container finished" podID="aa7ab277-af83-484c-917b-5398fbb64bcd" containerID="ab05544afde008e17bd8949b6444b499b86eab4cedc3487c997c6113f6b4d231" exitCode=0 Oct 03 14:10:34 crc kubenswrapper[4578]: I1003 14:10:34.196335 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z7wfg" event={"ID":"aa7ab277-af83-484c-917b-5398fbb64bcd","Type":"ContainerDied","Data":"ab05544afde008e17bd8949b6444b499b86eab4cedc3487c997c6113f6b4d231"} Oct 03 14:10:34 crc kubenswrapper[4578]: I1003 14:10:34.196362 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z7wfg" event={"ID":"aa7ab277-af83-484c-917b-5398fbb64bcd","Type":"ContainerDied","Data":"3c51741347aacfd1d659fc92444f0fc66a5b2d7bc41e65baa0d754ef422afbe8"} Oct 03 14:10:34 crc kubenswrapper[4578]: I1003 14:10:34.196378 4578 scope.go:117] "RemoveContainer" containerID="ab05544afde008e17bd8949b6444b499b86eab4cedc3487c997c6113f6b4d231" Oct 03 14:10:34 crc kubenswrapper[4578]: I1003 14:10:34.196391 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z7wfg" Oct 03 14:10:34 crc kubenswrapper[4578]: I1003 14:10:34.243756 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z7wfg"] Oct 03 14:10:34 crc kubenswrapper[4578]: I1003 14:10:34.251234 4578 scope.go:117] "RemoveContainer" containerID="191d40d7c13db3659433f41ef668c794150fc4bc36d4f113095774871130c999" Oct 03 14:10:34 crc kubenswrapper[4578]: I1003 14:10:34.256870 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-z7wfg"] Oct 03 14:10:34 crc kubenswrapper[4578]: I1003 14:10:34.275467 4578 scope.go:117] "RemoveContainer" containerID="fde50f866088e5c235bd47c4e4ae9fc3bf1028aa6b27ed0827aa63d14d181d12" Oct 03 14:10:34 crc kubenswrapper[4578]: I1003 14:10:34.331214 4578 scope.go:117] "RemoveContainer" containerID="ab05544afde008e17bd8949b6444b499b86eab4cedc3487c997c6113f6b4d231" Oct 03 14:10:34 crc kubenswrapper[4578]: E1003 14:10:34.331710 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab05544afde008e17bd8949b6444b499b86eab4cedc3487c997c6113f6b4d231\": container with ID starting with ab05544afde008e17bd8949b6444b499b86eab4cedc3487c997c6113f6b4d231 not found: ID does not exist" containerID="ab05544afde008e17bd8949b6444b499b86eab4cedc3487c997c6113f6b4d231" Oct 03 14:10:34 crc kubenswrapper[4578]: I1003 14:10:34.331761 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab05544afde008e17bd8949b6444b499b86eab4cedc3487c997c6113f6b4d231"} err="failed to get container status \"ab05544afde008e17bd8949b6444b499b86eab4cedc3487c997c6113f6b4d231\": rpc error: code = NotFound desc = could not find container \"ab05544afde008e17bd8949b6444b499b86eab4cedc3487c997c6113f6b4d231\": container with ID starting with ab05544afde008e17bd8949b6444b499b86eab4cedc3487c997c6113f6b4d231 not found: ID does not exist" Oct 03 14:10:34 crc kubenswrapper[4578]: I1003 14:10:34.331793 4578 scope.go:117] "RemoveContainer" containerID="191d40d7c13db3659433f41ef668c794150fc4bc36d4f113095774871130c999" Oct 03 14:10:34 crc kubenswrapper[4578]: E1003 14:10:34.332075 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"191d40d7c13db3659433f41ef668c794150fc4bc36d4f113095774871130c999\": container with ID starting with 191d40d7c13db3659433f41ef668c794150fc4bc36d4f113095774871130c999 not found: ID does not exist" containerID="191d40d7c13db3659433f41ef668c794150fc4bc36d4f113095774871130c999" Oct 03 14:10:34 crc kubenswrapper[4578]: I1003 14:10:34.332128 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"191d40d7c13db3659433f41ef668c794150fc4bc36d4f113095774871130c999"} err="failed to get container status \"191d40d7c13db3659433f41ef668c794150fc4bc36d4f113095774871130c999\": rpc error: code = NotFound desc = could not find container \"191d40d7c13db3659433f41ef668c794150fc4bc36d4f113095774871130c999\": container with ID starting with 191d40d7c13db3659433f41ef668c794150fc4bc36d4f113095774871130c999 not found: ID does not exist" Oct 03 14:10:34 crc kubenswrapper[4578]: I1003 14:10:34.332146 4578 scope.go:117] "RemoveContainer" containerID="fde50f866088e5c235bd47c4e4ae9fc3bf1028aa6b27ed0827aa63d14d181d12" Oct 03 14:10:34 crc kubenswrapper[4578]: E1003 14:10:34.332496 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fde50f866088e5c235bd47c4e4ae9fc3bf1028aa6b27ed0827aa63d14d181d12\": container with ID starting with fde50f866088e5c235bd47c4e4ae9fc3bf1028aa6b27ed0827aa63d14d181d12 not found: ID does not exist" containerID="fde50f866088e5c235bd47c4e4ae9fc3bf1028aa6b27ed0827aa63d14d181d12" Oct 03 14:10:34 crc kubenswrapper[4578]: I1003 14:10:34.332542 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fde50f866088e5c235bd47c4e4ae9fc3bf1028aa6b27ed0827aa63d14d181d12"} err="failed to get container status \"fde50f866088e5c235bd47c4e4ae9fc3bf1028aa6b27ed0827aa63d14d181d12\": rpc error: code = NotFound desc = could not find container \"fde50f866088e5c235bd47c4e4ae9fc3bf1028aa6b27ed0827aa63d14d181d12\": container with ID starting with fde50f866088e5c235bd47c4e4ae9fc3bf1028aa6b27ed0827aa63d14d181d12 not found: ID does not exist" Oct 03 14:10:34 crc kubenswrapper[4578]: I1003 14:10:34.919893 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa7ab277-af83-484c-917b-5398fbb64bcd" path="/var/lib/kubelet/pods/aa7ab277-af83-484c-917b-5398fbb64bcd/volumes" Oct 03 14:10:35 crc kubenswrapper[4578]: I1003 14:10:35.091604 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:10:35 crc kubenswrapper[4578]: I1003 14:10:35.091723 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:10:35 crc kubenswrapper[4578]: I1003 14:10:35.091791 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 14:10:35 crc kubenswrapper[4578]: I1003 14:10:35.093063 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d75edac35d3e8d13aacc39445c65e996b89e3d449792b96e3b020bb14e4e8800"} pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 14:10:35 crc kubenswrapper[4578]: I1003 14:10:35.093212 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" containerID="cri-o://d75edac35d3e8d13aacc39445c65e996b89e3d449792b96e3b020bb14e4e8800" gracePeriod=600 Oct 03 14:10:36 crc kubenswrapper[4578]: I1003 14:10:36.225086 4578 generic.go:334] "Generic (PLEG): container finished" podID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerID="d75edac35d3e8d13aacc39445c65e996b89e3d449792b96e3b020bb14e4e8800" exitCode=0 Oct 03 14:10:36 crc kubenswrapper[4578]: I1003 14:10:36.225195 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerDied","Data":"d75edac35d3e8d13aacc39445c65e996b89e3d449792b96e3b020bb14e4e8800"} Oct 03 14:10:36 crc kubenswrapper[4578]: I1003 14:10:36.225717 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577"} Oct 03 14:10:36 crc kubenswrapper[4578]: I1003 14:10:36.225745 4578 scope.go:117] "RemoveContainer" containerID="8c4336c0ada999c138d24a58f16ada9e3e8f9f9941d45e48883924db81c5e3c6" Oct 03 14:11:32 crc kubenswrapper[4578]: I1003 14:11:32.699428 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q9sjj"] Oct 03 14:11:32 crc kubenswrapper[4578]: E1003 14:11:32.701644 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa7ab277-af83-484c-917b-5398fbb64bcd" containerName="registry-server" Oct 03 14:11:32 crc kubenswrapper[4578]: I1003 14:11:32.701726 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa7ab277-af83-484c-917b-5398fbb64bcd" containerName="registry-server" Oct 03 14:11:32 crc kubenswrapper[4578]: E1003 14:11:32.701785 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa7ab277-af83-484c-917b-5398fbb64bcd" containerName="extract-utilities" Oct 03 14:11:32 crc kubenswrapper[4578]: I1003 14:11:32.701836 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa7ab277-af83-484c-917b-5398fbb64bcd" containerName="extract-utilities" Oct 03 14:11:32 crc kubenswrapper[4578]: E1003 14:11:32.701927 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa7ab277-af83-484c-917b-5398fbb64bcd" containerName="extract-content" Oct 03 14:11:32 crc kubenswrapper[4578]: I1003 14:11:32.701991 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa7ab277-af83-484c-917b-5398fbb64bcd" containerName="extract-content" Oct 03 14:11:32 crc kubenswrapper[4578]: I1003 14:11:32.702245 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa7ab277-af83-484c-917b-5398fbb64bcd" containerName="registry-server" Oct 03 14:11:32 crc kubenswrapper[4578]: I1003 14:11:32.705349 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q9sjj" Oct 03 14:11:32 crc kubenswrapper[4578]: I1003 14:11:32.717670 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q9sjj"] Oct 03 14:11:32 crc kubenswrapper[4578]: I1003 14:11:32.875554 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm8vh\" (UniqueName: \"kubernetes.io/projected/a577ad72-330b-4323-870d-22fc986335ce-kube-api-access-fm8vh\") pod \"redhat-operators-q9sjj\" (UID: \"a577ad72-330b-4323-870d-22fc986335ce\") " pod="openshift-marketplace/redhat-operators-q9sjj" Oct 03 14:11:32 crc kubenswrapper[4578]: I1003 14:11:32.875671 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a577ad72-330b-4323-870d-22fc986335ce-utilities\") pod \"redhat-operators-q9sjj\" (UID: \"a577ad72-330b-4323-870d-22fc986335ce\") " pod="openshift-marketplace/redhat-operators-q9sjj" Oct 03 14:11:32 crc kubenswrapper[4578]: I1003 14:11:32.875787 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a577ad72-330b-4323-870d-22fc986335ce-catalog-content\") pod \"redhat-operators-q9sjj\" (UID: \"a577ad72-330b-4323-870d-22fc986335ce\") " pod="openshift-marketplace/redhat-operators-q9sjj" Oct 03 14:11:32 crc kubenswrapper[4578]: I1003 14:11:32.978147 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm8vh\" (UniqueName: \"kubernetes.io/projected/a577ad72-330b-4323-870d-22fc986335ce-kube-api-access-fm8vh\") pod \"redhat-operators-q9sjj\" (UID: \"a577ad72-330b-4323-870d-22fc986335ce\") " pod="openshift-marketplace/redhat-operators-q9sjj" Oct 03 14:11:32 crc kubenswrapper[4578]: I1003 14:11:32.978321 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a577ad72-330b-4323-870d-22fc986335ce-utilities\") pod \"redhat-operators-q9sjj\" (UID: \"a577ad72-330b-4323-870d-22fc986335ce\") " pod="openshift-marketplace/redhat-operators-q9sjj" Oct 03 14:11:32 crc kubenswrapper[4578]: I1003 14:11:32.978509 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a577ad72-330b-4323-870d-22fc986335ce-catalog-content\") pod \"redhat-operators-q9sjj\" (UID: \"a577ad72-330b-4323-870d-22fc986335ce\") " pod="openshift-marketplace/redhat-operators-q9sjj" Oct 03 14:11:32 crc kubenswrapper[4578]: I1003 14:11:32.979032 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a577ad72-330b-4323-870d-22fc986335ce-catalog-content\") pod \"redhat-operators-q9sjj\" (UID: \"a577ad72-330b-4323-870d-22fc986335ce\") " pod="openshift-marketplace/redhat-operators-q9sjj" Oct 03 14:11:32 crc kubenswrapper[4578]: I1003 14:11:32.979113 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a577ad72-330b-4323-870d-22fc986335ce-utilities\") pod \"redhat-operators-q9sjj\" (UID: \"a577ad72-330b-4323-870d-22fc986335ce\") " pod="openshift-marketplace/redhat-operators-q9sjj" Oct 03 14:11:32 crc kubenswrapper[4578]: I1003 14:11:32.996667 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm8vh\" (UniqueName: \"kubernetes.io/projected/a577ad72-330b-4323-870d-22fc986335ce-kube-api-access-fm8vh\") pod \"redhat-operators-q9sjj\" (UID: \"a577ad72-330b-4323-870d-22fc986335ce\") " pod="openshift-marketplace/redhat-operators-q9sjj" Oct 03 14:11:33 crc kubenswrapper[4578]: I1003 14:11:33.031583 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q9sjj" Oct 03 14:11:33 crc kubenswrapper[4578]: I1003 14:11:33.510333 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q9sjj"] Oct 03 14:11:33 crc kubenswrapper[4578]: I1003 14:11:33.775123 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9sjj" event={"ID":"a577ad72-330b-4323-870d-22fc986335ce","Type":"ContainerStarted","Data":"46b8d1f261c2f765c580ae0e0482cde00085dda1bb3be446a0fab38fa0176789"} Oct 03 14:11:33 crc kubenswrapper[4578]: I1003 14:11:33.776028 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9sjj" event={"ID":"a577ad72-330b-4323-870d-22fc986335ce","Type":"ContainerStarted","Data":"1eee34d911a71a2d6f24e1077b445420616560552a5678185ed4110dfe3ead18"} Oct 03 14:11:34 crc kubenswrapper[4578]: I1003 14:11:34.782753 4578 generic.go:334] "Generic (PLEG): container finished" podID="a577ad72-330b-4323-870d-22fc986335ce" containerID="46b8d1f261c2f765c580ae0e0482cde00085dda1bb3be446a0fab38fa0176789" exitCode=0 Oct 03 14:11:34 crc kubenswrapper[4578]: I1003 14:11:34.783033 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9sjj" event={"ID":"a577ad72-330b-4323-870d-22fc986335ce","Type":"ContainerDied","Data":"46b8d1f261c2f765c580ae0e0482cde00085dda1bb3be446a0fab38fa0176789"} Oct 03 14:11:36 crc kubenswrapper[4578]: I1003 14:11:36.824389 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9sjj" event={"ID":"a577ad72-330b-4323-870d-22fc986335ce","Type":"ContainerStarted","Data":"c39c6395495c172dd4ccce3d5ee4dac88cf6669c1dc7e13ad9d764730b039f45"} Oct 03 14:11:38 crc kubenswrapper[4578]: I1003 14:11:38.847742 4578 generic.go:334] "Generic (PLEG): container finished" podID="a577ad72-330b-4323-870d-22fc986335ce" containerID="c39c6395495c172dd4ccce3d5ee4dac88cf6669c1dc7e13ad9d764730b039f45" exitCode=0 Oct 03 14:11:38 crc kubenswrapper[4578]: I1003 14:11:38.847745 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9sjj" event={"ID":"a577ad72-330b-4323-870d-22fc986335ce","Type":"ContainerDied","Data":"c39c6395495c172dd4ccce3d5ee4dac88cf6669c1dc7e13ad9d764730b039f45"} Oct 03 14:11:39 crc kubenswrapper[4578]: I1003 14:11:39.859071 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9sjj" event={"ID":"a577ad72-330b-4323-870d-22fc986335ce","Type":"ContainerStarted","Data":"c30c6d0bc5c937395d68aa9e343dffbb34798c64dc2a12d5d51ca599ff2e1e73"} Oct 03 14:11:39 crc kubenswrapper[4578]: I1003 14:11:39.882897 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q9sjj" podStartSLOduration=3.325311855 podStartE2EDuration="7.882875444s" podCreationTimestamp="2025-10-03 14:11:32 +0000 UTC" firstStartedPulling="2025-10-03 14:11:34.784562497 +0000 UTC m=+4830.583034691" lastFinishedPulling="2025-10-03 14:11:39.342126096 +0000 UTC m=+4835.140598280" observedRunningTime="2025-10-03 14:11:39.878185308 +0000 UTC m=+4835.676657502" watchObservedRunningTime="2025-10-03 14:11:39.882875444 +0000 UTC m=+4835.681347628" Oct 03 14:11:43 crc kubenswrapper[4578]: I1003 14:11:43.032704 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q9sjj" Oct 03 14:11:43 crc kubenswrapper[4578]: I1003 14:11:43.033088 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q9sjj" Oct 03 14:11:44 crc kubenswrapper[4578]: I1003 14:11:44.078720 4578 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-q9sjj" podUID="a577ad72-330b-4323-870d-22fc986335ce" containerName="registry-server" probeResult="failure" output=< Oct 03 14:11:44 crc kubenswrapper[4578]: timeout: failed to connect service ":50051" within 1s Oct 03 14:11:44 crc kubenswrapper[4578]: > Oct 03 14:11:53 crc kubenswrapper[4578]: I1003 14:11:53.079604 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q9sjj" Oct 03 14:11:53 crc kubenswrapper[4578]: I1003 14:11:53.127533 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q9sjj" Oct 03 14:11:53 crc kubenswrapper[4578]: I1003 14:11:53.313969 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q9sjj"] Oct 03 14:11:55 crc kubenswrapper[4578]: I1003 14:11:55.021852 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q9sjj" podUID="a577ad72-330b-4323-870d-22fc986335ce" containerName="registry-server" containerID="cri-o://c30c6d0bc5c937395d68aa9e343dffbb34798c64dc2a12d5d51ca599ff2e1e73" gracePeriod=2 Oct 03 14:11:55 crc kubenswrapper[4578]: I1003 14:11:55.782909 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q9sjj" Oct 03 14:11:55 crc kubenswrapper[4578]: I1003 14:11:55.879908 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a577ad72-330b-4323-870d-22fc986335ce-utilities\") pod \"a577ad72-330b-4323-870d-22fc986335ce\" (UID: \"a577ad72-330b-4323-870d-22fc986335ce\") " Oct 03 14:11:55 crc kubenswrapper[4578]: I1003 14:11:55.879947 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a577ad72-330b-4323-870d-22fc986335ce-catalog-content\") pod \"a577ad72-330b-4323-870d-22fc986335ce\" (UID: \"a577ad72-330b-4323-870d-22fc986335ce\") " Oct 03 14:11:55 crc kubenswrapper[4578]: I1003 14:11:55.879975 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fm8vh\" (UniqueName: \"kubernetes.io/projected/a577ad72-330b-4323-870d-22fc986335ce-kube-api-access-fm8vh\") pod \"a577ad72-330b-4323-870d-22fc986335ce\" (UID: \"a577ad72-330b-4323-870d-22fc986335ce\") " Oct 03 14:11:55 crc kubenswrapper[4578]: I1003 14:11:55.881379 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a577ad72-330b-4323-870d-22fc986335ce-utilities" (OuterVolumeSpecName: "utilities") pod "a577ad72-330b-4323-870d-22fc986335ce" (UID: "a577ad72-330b-4323-870d-22fc986335ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:11:55 crc kubenswrapper[4578]: I1003 14:11:55.886811 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a577ad72-330b-4323-870d-22fc986335ce-kube-api-access-fm8vh" (OuterVolumeSpecName: "kube-api-access-fm8vh") pod "a577ad72-330b-4323-870d-22fc986335ce" (UID: "a577ad72-330b-4323-870d-22fc986335ce"). InnerVolumeSpecName "kube-api-access-fm8vh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:11:55 crc kubenswrapper[4578]: I1003 14:11:55.967184 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a577ad72-330b-4323-870d-22fc986335ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a577ad72-330b-4323-870d-22fc986335ce" (UID: "a577ad72-330b-4323-870d-22fc986335ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:11:55 crc kubenswrapper[4578]: I1003 14:11:55.981558 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a577ad72-330b-4323-870d-22fc986335ce-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:11:55 crc kubenswrapper[4578]: I1003 14:11:55.981591 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a577ad72-330b-4323-870d-22fc986335ce-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:11:55 crc kubenswrapper[4578]: I1003 14:11:55.981603 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fm8vh\" (UniqueName: \"kubernetes.io/projected/a577ad72-330b-4323-870d-22fc986335ce-kube-api-access-fm8vh\") on node \"crc\" DevicePath \"\"" Oct 03 14:11:56 crc kubenswrapper[4578]: I1003 14:11:56.033425 4578 generic.go:334] "Generic (PLEG): container finished" podID="a577ad72-330b-4323-870d-22fc986335ce" containerID="c30c6d0bc5c937395d68aa9e343dffbb34798c64dc2a12d5d51ca599ff2e1e73" exitCode=0 Oct 03 14:11:56 crc kubenswrapper[4578]: I1003 14:11:56.033477 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9sjj" event={"ID":"a577ad72-330b-4323-870d-22fc986335ce","Type":"ContainerDied","Data":"c30c6d0bc5c937395d68aa9e343dffbb34798c64dc2a12d5d51ca599ff2e1e73"} Oct 03 14:11:56 crc kubenswrapper[4578]: I1003 14:11:56.033508 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q9sjj" event={"ID":"a577ad72-330b-4323-870d-22fc986335ce","Type":"ContainerDied","Data":"1eee34d911a71a2d6f24e1077b445420616560552a5678185ed4110dfe3ead18"} Oct 03 14:11:56 crc kubenswrapper[4578]: I1003 14:11:56.033529 4578 scope.go:117] "RemoveContainer" containerID="c30c6d0bc5c937395d68aa9e343dffbb34798c64dc2a12d5d51ca599ff2e1e73" Oct 03 14:11:56 crc kubenswrapper[4578]: I1003 14:11:56.033571 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q9sjj" Oct 03 14:11:56 crc kubenswrapper[4578]: I1003 14:11:56.082234 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q9sjj"] Oct 03 14:11:56 crc kubenswrapper[4578]: I1003 14:11:56.091340 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q9sjj"] Oct 03 14:11:56 crc kubenswrapper[4578]: I1003 14:11:56.096911 4578 scope.go:117] "RemoveContainer" containerID="c39c6395495c172dd4ccce3d5ee4dac88cf6669c1dc7e13ad9d764730b039f45" Oct 03 14:11:56 crc kubenswrapper[4578]: I1003 14:11:56.117943 4578 scope.go:117] "RemoveContainer" containerID="46b8d1f261c2f765c580ae0e0482cde00085dda1bb3be446a0fab38fa0176789" Oct 03 14:11:56 crc kubenswrapper[4578]: I1003 14:11:56.161013 4578 scope.go:117] "RemoveContainer" containerID="c30c6d0bc5c937395d68aa9e343dffbb34798c64dc2a12d5d51ca599ff2e1e73" Oct 03 14:11:56 crc kubenswrapper[4578]: E1003 14:11:56.161492 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c30c6d0bc5c937395d68aa9e343dffbb34798c64dc2a12d5d51ca599ff2e1e73\": container with ID starting with c30c6d0bc5c937395d68aa9e343dffbb34798c64dc2a12d5d51ca599ff2e1e73 not found: ID does not exist" containerID="c30c6d0bc5c937395d68aa9e343dffbb34798c64dc2a12d5d51ca599ff2e1e73" Oct 03 14:11:56 crc kubenswrapper[4578]: I1003 14:11:56.161528 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c30c6d0bc5c937395d68aa9e343dffbb34798c64dc2a12d5d51ca599ff2e1e73"} err="failed to get container status \"c30c6d0bc5c937395d68aa9e343dffbb34798c64dc2a12d5d51ca599ff2e1e73\": rpc error: code = NotFound desc = could not find container \"c30c6d0bc5c937395d68aa9e343dffbb34798c64dc2a12d5d51ca599ff2e1e73\": container with ID starting with c30c6d0bc5c937395d68aa9e343dffbb34798c64dc2a12d5d51ca599ff2e1e73 not found: ID does not exist" Oct 03 14:11:56 crc kubenswrapper[4578]: I1003 14:11:56.161554 4578 scope.go:117] "RemoveContainer" containerID="c39c6395495c172dd4ccce3d5ee4dac88cf6669c1dc7e13ad9d764730b039f45" Oct 03 14:11:56 crc kubenswrapper[4578]: E1003 14:11:56.161994 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c39c6395495c172dd4ccce3d5ee4dac88cf6669c1dc7e13ad9d764730b039f45\": container with ID starting with c39c6395495c172dd4ccce3d5ee4dac88cf6669c1dc7e13ad9d764730b039f45 not found: ID does not exist" containerID="c39c6395495c172dd4ccce3d5ee4dac88cf6669c1dc7e13ad9d764730b039f45" Oct 03 14:11:56 crc kubenswrapper[4578]: I1003 14:11:56.162020 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c39c6395495c172dd4ccce3d5ee4dac88cf6669c1dc7e13ad9d764730b039f45"} err="failed to get container status \"c39c6395495c172dd4ccce3d5ee4dac88cf6669c1dc7e13ad9d764730b039f45\": rpc error: code = NotFound desc = could not find container \"c39c6395495c172dd4ccce3d5ee4dac88cf6669c1dc7e13ad9d764730b039f45\": container with ID starting with c39c6395495c172dd4ccce3d5ee4dac88cf6669c1dc7e13ad9d764730b039f45 not found: ID does not exist" Oct 03 14:11:56 crc kubenswrapper[4578]: I1003 14:11:56.162041 4578 scope.go:117] "RemoveContainer" containerID="46b8d1f261c2f765c580ae0e0482cde00085dda1bb3be446a0fab38fa0176789" Oct 03 14:11:56 crc kubenswrapper[4578]: E1003 14:11:56.162598 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46b8d1f261c2f765c580ae0e0482cde00085dda1bb3be446a0fab38fa0176789\": container with ID starting with 46b8d1f261c2f765c580ae0e0482cde00085dda1bb3be446a0fab38fa0176789 not found: ID does not exist" containerID="46b8d1f261c2f765c580ae0e0482cde00085dda1bb3be446a0fab38fa0176789" Oct 03 14:11:56 crc kubenswrapper[4578]: I1003 14:11:56.162718 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46b8d1f261c2f765c580ae0e0482cde00085dda1bb3be446a0fab38fa0176789"} err="failed to get container status \"46b8d1f261c2f765c580ae0e0482cde00085dda1bb3be446a0fab38fa0176789\": rpc error: code = NotFound desc = could not find container \"46b8d1f261c2f765c580ae0e0482cde00085dda1bb3be446a0fab38fa0176789\": container with ID starting with 46b8d1f261c2f765c580ae0e0482cde00085dda1bb3be446a0fab38fa0176789 not found: ID does not exist" Oct 03 14:11:56 crc kubenswrapper[4578]: I1003 14:11:56.918288 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a577ad72-330b-4323-870d-22fc986335ce" path="/var/lib/kubelet/pods/a577ad72-330b-4323-870d-22fc986335ce/volumes" Oct 03 14:12:35 crc kubenswrapper[4578]: I1003 14:12:35.091963 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:12:35 crc kubenswrapper[4578]: I1003 14:12:35.092559 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:13:05 crc kubenswrapper[4578]: I1003 14:13:05.091848 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:13:05 crc kubenswrapper[4578]: I1003 14:13:05.092347 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:13:35 crc kubenswrapper[4578]: I1003 14:13:35.091670 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:13:35 crc kubenswrapper[4578]: I1003 14:13:35.092205 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:13:35 crc kubenswrapper[4578]: I1003 14:13:35.092257 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 14:13:35 crc kubenswrapper[4578]: I1003 14:13:35.093215 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577"} pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 14:13:35 crc kubenswrapper[4578]: I1003 14:13:35.093284 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" containerID="cri-o://fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" gracePeriod=600 Oct 03 14:13:35 crc kubenswrapper[4578]: E1003 14:13:35.216132 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:13:35 crc kubenswrapper[4578]: I1003 14:13:35.987526 4578 generic.go:334] "Generic (PLEG): container finished" podID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" exitCode=0 Oct 03 14:13:35 crc kubenswrapper[4578]: I1003 14:13:35.987597 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerDied","Data":"fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577"} Oct 03 14:13:35 crc kubenswrapper[4578]: I1003 14:13:35.987661 4578 scope.go:117] "RemoveContainer" containerID="d75edac35d3e8d13aacc39445c65e996b89e3d449792b96e3b020bb14e4e8800" Oct 03 14:13:35 crc kubenswrapper[4578]: I1003 14:13:35.988376 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:13:35 crc kubenswrapper[4578]: E1003 14:13:35.988612 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:13:48 crc kubenswrapper[4578]: I1003 14:13:48.908975 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:13:48 crc kubenswrapper[4578]: E1003 14:13:48.909783 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:14:02 crc kubenswrapper[4578]: I1003 14:14:02.908985 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:14:02 crc kubenswrapper[4578]: E1003 14:14:02.909744 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:14:16 crc kubenswrapper[4578]: I1003 14:14:16.909977 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:14:16 crc kubenswrapper[4578]: E1003 14:14:16.911087 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:14:30 crc kubenswrapper[4578]: I1003 14:14:30.910127 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:14:30 crc kubenswrapper[4578]: E1003 14:14:30.911067 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:14:43 crc kubenswrapper[4578]: I1003 14:14:43.909665 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:14:43 crc kubenswrapper[4578]: E1003 14:14:43.910480 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:14:54 crc kubenswrapper[4578]: I1003 14:14:54.917720 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:14:54 crc kubenswrapper[4578]: E1003 14:14:54.918608 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:15:00 crc kubenswrapper[4578]: I1003 14:15:00.178526 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325015-v2tf2"] Oct 03 14:15:00 crc kubenswrapper[4578]: E1003 14:15:00.179710 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a577ad72-330b-4323-870d-22fc986335ce" containerName="extract-content" Oct 03 14:15:00 crc kubenswrapper[4578]: I1003 14:15:00.179730 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a577ad72-330b-4323-870d-22fc986335ce" containerName="extract-content" Oct 03 14:15:00 crc kubenswrapper[4578]: E1003 14:15:00.179786 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a577ad72-330b-4323-870d-22fc986335ce" containerName="registry-server" Oct 03 14:15:00 crc kubenswrapper[4578]: I1003 14:15:00.179798 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a577ad72-330b-4323-870d-22fc986335ce" containerName="registry-server" Oct 03 14:15:00 crc kubenswrapper[4578]: E1003 14:15:00.179834 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a577ad72-330b-4323-870d-22fc986335ce" containerName="extract-utilities" Oct 03 14:15:00 crc kubenswrapper[4578]: I1003 14:15:00.179847 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="a577ad72-330b-4323-870d-22fc986335ce" containerName="extract-utilities" Oct 03 14:15:00 crc kubenswrapper[4578]: I1003 14:15:00.180170 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="a577ad72-330b-4323-870d-22fc986335ce" containerName="registry-server" Oct 03 14:15:00 crc kubenswrapper[4578]: I1003 14:15:00.181199 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-v2tf2" Oct 03 14:15:00 crc kubenswrapper[4578]: I1003 14:15:00.186586 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 03 14:15:00 crc kubenswrapper[4578]: I1003 14:15:00.187757 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 03 14:15:00 crc kubenswrapper[4578]: I1003 14:15:00.196989 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325015-v2tf2"] Oct 03 14:15:00 crc kubenswrapper[4578]: I1003 14:15:00.319055 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1068b648-059c-4571-84bc-d76adb4ca57a-config-volume\") pod \"collect-profiles-29325015-v2tf2\" (UID: \"1068b648-059c-4571-84bc-d76adb4ca57a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-v2tf2" Oct 03 14:15:00 crc kubenswrapper[4578]: I1003 14:15:00.320663 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pt9p\" (UniqueName: \"kubernetes.io/projected/1068b648-059c-4571-84bc-d76adb4ca57a-kube-api-access-5pt9p\") pod \"collect-profiles-29325015-v2tf2\" (UID: \"1068b648-059c-4571-84bc-d76adb4ca57a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-v2tf2" Oct 03 14:15:00 crc kubenswrapper[4578]: I1003 14:15:00.320760 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1068b648-059c-4571-84bc-d76adb4ca57a-secret-volume\") pod \"collect-profiles-29325015-v2tf2\" (UID: \"1068b648-059c-4571-84bc-d76adb4ca57a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-v2tf2" Oct 03 14:15:00 crc kubenswrapper[4578]: I1003 14:15:00.422226 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1068b648-059c-4571-84bc-d76adb4ca57a-secret-volume\") pod \"collect-profiles-29325015-v2tf2\" (UID: \"1068b648-059c-4571-84bc-d76adb4ca57a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-v2tf2" Oct 03 14:15:00 crc kubenswrapper[4578]: I1003 14:15:00.422346 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1068b648-059c-4571-84bc-d76adb4ca57a-config-volume\") pod \"collect-profiles-29325015-v2tf2\" (UID: \"1068b648-059c-4571-84bc-d76adb4ca57a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-v2tf2" Oct 03 14:15:00 crc kubenswrapper[4578]: I1003 14:15:00.422413 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pt9p\" (UniqueName: \"kubernetes.io/projected/1068b648-059c-4571-84bc-d76adb4ca57a-kube-api-access-5pt9p\") pod \"collect-profiles-29325015-v2tf2\" (UID: \"1068b648-059c-4571-84bc-d76adb4ca57a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-v2tf2" Oct 03 14:15:00 crc kubenswrapper[4578]: I1003 14:15:00.423484 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1068b648-059c-4571-84bc-d76adb4ca57a-config-volume\") pod \"collect-profiles-29325015-v2tf2\" (UID: \"1068b648-059c-4571-84bc-d76adb4ca57a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-v2tf2" Oct 03 14:15:00 crc kubenswrapper[4578]: I1003 14:15:00.438871 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1068b648-059c-4571-84bc-d76adb4ca57a-secret-volume\") pod \"collect-profiles-29325015-v2tf2\" (UID: \"1068b648-059c-4571-84bc-d76adb4ca57a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-v2tf2" Oct 03 14:15:00 crc kubenswrapper[4578]: I1003 14:15:00.439214 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pt9p\" (UniqueName: \"kubernetes.io/projected/1068b648-059c-4571-84bc-d76adb4ca57a-kube-api-access-5pt9p\") pod \"collect-profiles-29325015-v2tf2\" (UID: \"1068b648-059c-4571-84bc-d76adb4ca57a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-v2tf2" Oct 03 14:15:00 crc kubenswrapper[4578]: I1003 14:15:00.508285 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-v2tf2" Oct 03 14:15:00 crc kubenswrapper[4578]: I1003 14:15:00.945401 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325015-v2tf2"] Oct 03 14:15:01 crc kubenswrapper[4578]: I1003 14:15:01.827519 4578 generic.go:334] "Generic (PLEG): container finished" podID="1068b648-059c-4571-84bc-d76adb4ca57a" containerID="55b2808ac86e642d35f440b74bcf389b87c691cb304b7a8a199556fef622bf5a" exitCode=0 Oct 03 14:15:01 crc kubenswrapper[4578]: I1003 14:15:01.827577 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-v2tf2" event={"ID":"1068b648-059c-4571-84bc-d76adb4ca57a","Type":"ContainerDied","Data":"55b2808ac86e642d35f440b74bcf389b87c691cb304b7a8a199556fef622bf5a"} Oct 03 14:15:01 crc kubenswrapper[4578]: I1003 14:15:01.828176 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-v2tf2" event={"ID":"1068b648-059c-4571-84bc-d76adb4ca57a","Type":"ContainerStarted","Data":"ac6ad1e77da40a02eea04b7f86db96502a6b1807b582789425eb1db85c71fe59"} Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.291740 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-v2tf2" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.404957 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bjs65"] Oct 03 14:15:03 crc kubenswrapper[4578]: E1003 14:15:03.405460 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1068b648-059c-4571-84bc-d76adb4ca57a" containerName="collect-profiles" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.405483 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="1068b648-059c-4571-84bc-d76adb4ca57a" containerName="collect-profiles" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.405769 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="1068b648-059c-4571-84bc-d76adb4ca57a" containerName="collect-profiles" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.415690 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bjs65" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.429881 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bjs65"] Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.489524 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1068b648-059c-4571-84bc-d76adb4ca57a-config-volume\") pod \"1068b648-059c-4571-84bc-d76adb4ca57a\" (UID: \"1068b648-059c-4571-84bc-d76adb4ca57a\") " Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.489565 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pt9p\" (UniqueName: \"kubernetes.io/projected/1068b648-059c-4571-84bc-d76adb4ca57a-kube-api-access-5pt9p\") pod \"1068b648-059c-4571-84bc-d76adb4ca57a\" (UID: \"1068b648-059c-4571-84bc-d76adb4ca57a\") " Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.489741 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1068b648-059c-4571-84bc-d76adb4ca57a-secret-volume\") pod \"1068b648-059c-4571-84bc-d76adb4ca57a\" (UID: \"1068b648-059c-4571-84bc-d76adb4ca57a\") " Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.490306 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1068b648-059c-4571-84bc-d76adb4ca57a-config-volume" (OuterVolumeSpecName: "config-volume") pod "1068b648-059c-4571-84bc-d76adb4ca57a" (UID: "1068b648-059c-4571-84bc-d76adb4ca57a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.496587 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1068b648-059c-4571-84bc-d76adb4ca57a-kube-api-access-5pt9p" (OuterVolumeSpecName: "kube-api-access-5pt9p") pod "1068b648-059c-4571-84bc-d76adb4ca57a" (UID: "1068b648-059c-4571-84bc-d76adb4ca57a"). InnerVolumeSpecName "kube-api-access-5pt9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.496618 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1068b648-059c-4571-84bc-d76adb4ca57a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1068b648-059c-4571-84bc-d76adb4ca57a" (UID: "1068b648-059c-4571-84bc-d76adb4ca57a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.592267 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v4cr\" (UniqueName: \"kubernetes.io/projected/391d359a-b909-4e02-8c37-142072970c22-kube-api-access-4v4cr\") pod \"community-operators-bjs65\" (UID: \"391d359a-b909-4e02-8c37-142072970c22\") " pod="openshift-marketplace/community-operators-bjs65" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.592329 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/391d359a-b909-4e02-8c37-142072970c22-utilities\") pod \"community-operators-bjs65\" (UID: \"391d359a-b909-4e02-8c37-142072970c22\") " pod="openshift-marketplace/community-operators-bjs65" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.592549 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/391d359a-b909-4e02-8c37-142072970c22-catalog-content\") pod \"community-operators-bjs65\" (UID: \"391d359a-b909-4e02-8c37-142072970c22\") " pod="openshift-marketplace/community-operators-bjs65" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.592619 4578 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1068b648-059c-4571-84bc-d76adb4ca57a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.592662 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pt9p\" (UniqueName: \"kubernetes.io/projected/1068b648-059c-4571-84bc-d76adb4ca57a-kube-api-access-5pt9p\") on node \"crc\" DevicePath \"\"" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.592676 4578 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1068b648-059c-4571-84bc-d76adb4ca57a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.694262 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/391d359a-b909-4e02-8c37-142072970c22-catalog-content\") pod \"community-operators-bjs65\" (UID: \"391d359a-b909-4e02-8c37-142072970c22\") " pod="openshift-marketplace/community-operators-bjs65" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.694331 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v4cr\" (UniqueName: \"kubernetes.io/projected/391d359a-b909-4e02-8c37-142072970c22-kube-api-access-4v4cr\") pod \"community-operators-bjs65\" (UID: \"391d359a-b909-4e02-8c37-142072970c22\") " pod="openshift-marketplace/community-operators-bjs65" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.694362 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/391d359a-b909-4e02-8c37-142072970c22-utilities\") pod \"community-operators-bjs65\" (UID: \"391d359a-b909-4e02-8c37-142072970c22\") " pod="openshift-marketplace/community-operators-bjs65" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.695002 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/391d359a-b909-4e02-8c37-142072970c22-utilities\") pod \"community-operators-bjs65\" (UID: \"391d359a-b909-4e02-8c37-142072970c22\") " pod="openshift-marketplace/community-operators-bjs65" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.695128 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/391d359a-b909-4e02-8c37-142072970c22-catalog-content\") pod \"community-operators-bjs65\" (UID: \"391d359a-b909-4e02-8c37-142072970c22\") " pod="openshift-marketplace/community-operators-bjs65" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.717047 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v4cr\" (UniqueName: \"kubernetes.io/projected/391d359a-b909-4e02-8c37-142072970c22-kube-api-access-4v4cr\") pod \"community-operators-bjs65\" (UID: \"391d359a-b909-4e02-8c37-142072970c22\") " pod="openshift-marketplace/community-operators-bjs65" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.739964 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bjs65" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.872045 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-v2tf2" event={"ID":"1068b648-059c-4571-84bc-d76adb4ca57a","Type":"ContainerDied","Data":"ac6ad1e77da40a02eea04b7f86db96502a6b1807b582789425eb1db85c71fe59"} Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.872297 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac6ad1e77da40a02eea04b7f86db96502a6b1807b582789425eb1db85c71fe59" Oct 03 14:15:03 crc kubenswrapper[4578]: I1003 14:15:03.872303 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325015-v2tf2" Oct 03 14:15:04 crc kubenswrapper[4578]: I1003 14:15:04.314441 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bjs65"] Oct 03 14:15:04 crc kubenswrapper[4578]: W1003 14:15:04.319830 4578 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod391d359a_b909_4e02_8c37_142072970c22.slice/crio-a124844dadee424251dad52ead14536460a4c9e63afad3f890330be570477129 WatchSource:0}: Error finding container a124844dadee424251dad52ead14536460a4c9e63afad3f890330be570477129: Status 404 returned error can't find the container with id a124844dadee424251dad52ead14536460a4c9e63afad3f890330be570477129 Oct 03 14:15:04 crc kubenswrapper[4578]: I1003 14:15:04.381812 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l"] Oct 03 14:15:04 crc kubenswrapper[4578]: I1003 14:15:04.386746 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29324970-bd42l"] Oct 03 14:15:04 crc kubenswrapper[4578]: I1003 14:15:04.883729 4578 generic.go:334] "Generic (PLEG): container finished" podID="391d359a-b909-4e02-8c37-142072970c22" containerID="94504235b793cd0adece6abc8d064e56d9bd6eb80afce8c3c77b2fb7fc541a15" exitCode=0 Oct 03 14:15:04 crc kubenswrapper[4578]: I1003 14:15:04.883802 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bjs65" event={"ID":"391d359a-b909-4e02-8c37-142072970c22","Type":"ContainerDied","Data":"94504235b793cd0adece6abc8d064e56d9bd6eb80afce8c3c77b2fb7fc541a15"} Oct 03 14:15:04 crc kubenswrapper[4578]: I1003 14:15:04.884087 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bjs65" event={"ID":"391d359a-b909-4e02-8c37-142072970c22","Type":"ContainerStarted","Data":"a124844dadee424251dad52ead14536460a4c9e63afad3f890330be570477129"} Oct 03 14:15:04 crc kubenswrapper[4578]: I1003 14:15:04.928726 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c92e0eba-3ff4-404b-a8ab-064f5ae88a12" path="/var/lib/kubelet/pods/c92e0eba-3ff4-404b-a8ab-064f5ae88a12/volumes" Oct 03 14:15:06 crc kubenswrapper[4578]: I1003 14:15:06.919581 4578 generic.go:334] "Generic (PLEG): container finished" podID="391d359a-b909-4e02-8c37-142072970c22" containerID="e08ab537c3dd40d25131dfb6d6b5843eb734bd97c88efcfd19695dc8b7d201d2" exitCode=0 Oct 03 14:15:06 crc kubenswrapper[4578]: I1003 14:15:06.921529 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bjs65" event={"ID":"391d359a-b909-4e02-8c37-142072970c22","Type":"ContainerDied","Data":"e08ab537c3dd40d25131dfb6d6b5843eb734bd97c88efcfd19695dc8b7d201d2"} Oct 03 14:15:07 crc kubenswrapper[4578]: I1003 14:15:07.930376 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bjs65" event={"ID":"391d359a-b909-4e02-8c37-142072970c22","Type":"ContainerStarted","Data":"eced3ae13958936f14bdc126a8423ed7e7131c7590022465ea27e34db6fef80d"} Oct 03 14:15:07 crc kubenswrapper[4578]: I1003 14:15:07.952170 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bjs65" podStartSLOduration=2.516133432 podStartE2EDuration="4.952146501s" podCreationTimestamp="2025-10-03 14:15:03 +0000 UTC" firstStartedPulling="2025-10-03 14:15:04.886456574 +0000 UTC m=+5040.684928758" lastFinishedPulling="2025-10-03 14:15:07.322469643 +0000 UTC m=+5043.120941827" observedRunningTime="2025-10-03 14:15:07.947778155 +0000 UTC m=+5043.746250359" watchObservedRunningTime="2025-10-03 14:15:07.952146501 +0000 UTC m=+5043.750618695" Oct 03 14:15:08 crc kubenswrapper[4578]: I1003 14:15:08.909126 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:15:08 crc kubenswrapper[4578]: E1003 14:15:08.909980 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:15:10 crc kubenswrapper[4578]: I1003 14:15:10.690292 4578 scope.go:117] "RemoveContainer" containerID="5cd906a0cd1fb7de0767611f43fe9ff637558dda5c1e1255bbcb7df80532e249" Oct 03 14:15:13 crc kubenswrapper[4578]: I1003 14:15:13.740092 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bjs65" Oct 03 14:15:13 crc kubenswrapper[4578]: I1003 14:15:13.740742 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bjs65" Oct 03 14:15:13 crc kubenswrapper[4578]: I1003 14:15:13.792707 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bjs65" Oct 03 14:15:14 crc kubenswrapper[4578]: I1003 14:15:14.035881 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bjs65" Oct 03 14:15:14 crc kubenswrapper[4578]: I1003 14:15:14.090548 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bjs65"] Oct 03 14:15:16 crc kubenswrapper[4578]: I1003 14:15:16.005520 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bjs65" podUID="391d359a-b909-4e02-8c37-142072970c22" containerName="registry-server" containerID="cri-o://eced3ae13958936f14bdc126a8423ed7e7131c7590022465ea27e34db6fef80d" gracePeriod=2 Oct 03 14:15:16 crc kubenswrapper[4578]: I1003 14:15:16.445382 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bjs65" Oct 03 14:15:16 crc kubenswrapper[4578]: I1003 14:15:16.541891 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/391d359a-b909-4e02-8c37-142072970c22-catalog-content\") pod \"391d359a-b909-4e02-8c37-142072970c22\" (UID: \"391d359a-b909-4e02-8c37-142072970c22\") " Oct 03 14:15:16 crc kubenswrapper[4578]: I1003 14:15:16.542309 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4v4cr\" (UniqueName: \"kubernetes.io/projected/391d359a-b909-4e02-8c37-142072970c22-kube-api-access-4v4cr\") pod \"391d359a-b909-4e02-8c37-142072970c22\" (UID: \"391d359a-b909-4e02-8c37-142072970c22\") " Oct 03 14:15:16 crc kubenswrapper[4578]: I1003 14:15:16.542349 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/391d359a-b909-4e02-8c37-142072970c22-utilities\") pod \"391d359a-b909-4e02-8c37-142072970c22\" (UID: \"391d359a-b909-4e02-8c37-142072970c22\") " Oct 03 14:15:16 crc kubenswrapper[4578]: I1003 14:15:16.543490 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/391d359a-b909-4e02-8c37-142072970c22-utilities" (OuterVolumeSpecName: "utilities") pod "391d359a-b909-4e02-8c37-142072970c22" (UID: "391d359a-b909-4e02-8c37-142072970c22"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:15:16 crc kubenswrapper[4578]: I1003 14:15:16.551959 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/391d359a-b909-4e02-8c37-142072970c22-kube-api-access-4v4cr" (OuterVolumeSpecName: "kube-api-access-4v4cr") pod "391d359a-b909-4e02-8c37-142072970c22" (UID: "391d359a-b909-4e02-8c37-142072970c22"). InnerVolumeSpecName "kube-api-access-4v4cr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:15:16 crc kubenswrapper[4578]: I1003 14:15:16.644781 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4v4cr\" (UniqueName: \"kubernetes.io/projected/391d359a-b909-4e02-8c37-142072970c22-kube-api-access-4v4cr\") on node \"crc\" DevicePath \"\"" Oct 03 14:15:16 crc kubenswrapper[4578]: I1003 14:15:16.644815 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/391d359a-b909-4e02-8c37-142072970c22-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:15:17 crc kubenswrapper[4578]: I1003 14:15:17.018229 4578 generic.go:334] "Generic (PLEG): container finished" podID="391d359a-b909-4e02-8c37-142072970c22" containerID="eced3ae13958936f14bdc126a8423ed7e7131c7590022465ea27e34db6fef80d" exitCode=0 Oct 03 14:15:17 crc kubenswrapper[4578]: I1003 14:15:17.018300 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bjs65" event={"ID":"391d359a-b909-4e02-8c37-142072970c22","Type":"ContainerDied","Data":"eced3ae13958936f14bdc126a8423ed7e7131c7590022465ea27e34db6fef80d"} Oct 03 14:15:17 crc kubenswrapper[4578]: I1003 14:15:17.018322 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bjs65" Oct 03 14:15:17 crc kubenswrapper[4578]: I1003 14:15:17.018351 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bjs65" event={"ID":"391d359a-b909-4e02-8c37-142072970c22","Type":"ContainerDied","Data":"a124844dadee424251dad52ead14536460a4c9e63afad3f890330be570477129"} Oct 03 14:15:17 crc kubenswrapper[4578]: I1003 14:15:17.018381 4578 scope.go:117] "RemoveContainer" containerID="eced3ae13958936f14bdc126a8423ed7e7131c7590022465ea27e34db6fef80d" Oct 03 14:15:17 crc kubenswrapper[4578]: I1003 14:15:17.032908 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/391d359a-b909-4e02-8c37-142072970c22-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "391d359a-b909-4e02-8c37-142072970c22" (UID: "391d359a-b909-4e02-8c37-142072970c22"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:15:17 crc kubenswrapper[4578]: I1003 14:15:17.050145 4578 scope.go:117] "RemoveContainer" containerID="e08ab537c3dd40d25131dfb6d6b5843eb734bd97c88efcfd19695dc8b7d201d2" Oct 03 14:15:17 crc kubenswrapper[4578]: I1003 14:15:17.054881 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/391d359a-b909-4e02-8c37-142072970c22-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:15:17 crc kubenswrapper[4578]: I1003 14:15:17.070356 4578 scope.go:117] "RemoveContainer" containerID="94504235b793cd0adece6abc8d064e56d9bd6eb80afce8c3c77b2fb7fc541a15" Oct 03 14:15:17 crc kubenswrapper[4578]: I1003 14:15:17.113684 4578 scope.go:117] "RemoveContainer" containerID="eced3ae13958936f14bdc126a8423ed7e7131c7590022465ea27e34db6fef80d" Oct 03 14:15:17 crc kubenswrapper[4578]: E1003 14:15:17.114050 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eced3ae13958936f14bdc126a8423ed7e7131c7590022465ea27e34db6fef80d\": container with ID starting with eced3ae13958936f14bdc126a8423ed7e7131c7590022465ea27e34db6fef80d not found: ID does not exist" containerID="eced3ae13958936f14bdc126a8423ed7e7131c7590022465ea27e34db6fef80d" Oct 03 14:15:17 crc kubenswrapper[4578]: I1003 14:15:17.114086 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eced3ae13958936f14bdc126a8423ed7e7131c7590022465ea27e34db6fef80d"} err="failed to get container status \"eced3ae13958936f14bdc126a8423ed7e7131c7590022465ea27e34db6fef80d\": rpc error: code = NotFound desc = could not find container \"eced3ae13958936f14bdc126a8423ed7e7131c7590022465ea27e34db6fef80d\": container with ID starting with eced3ae13958936f14bdc126a8423ed7e7131c7590022465ea27e34db6fef80d not found: ID does not exist" Oct 03 14:15:17 crc kubenswrapper[4578]: I1003 14:15:17.114109 4578 scope.go:117] "RemoveContainer" containerID="e08ab537c3dd40d25131dfb6d6b5843eb734bd97c88efcfd19695dc8b7d201d2" Oct 03 14:15:17 crc kubenswrapper[4578]: E1003 14:15:17.114385 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e08ab537c3dd40d25131dfb6d6b5843eb734bd97c88efcfd19695dc8b7d201d2\": container with ID starting with e08ab537c3dd40d25131dfb6d6b5843eb734bd97c88efcfd19695dc8b7d201d2 not found: ID does not exist" containerID="e08ab537c3dd40d25131dfb6d6b5843eb734bd97c88efcfd19695dc8b7d201d2" Oct 03 14:15:17 crc kubenswrapper[4578]: I1003 14:15:17.114410 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e08ab537c3dd40d25131dfb6d6b5843eb734bd97c88efcfd19695dc8b7d201d2"} err="failed to get container status \"e08ab537c3dd40d25131dfb6d6b5843eb734bd97c88efcfd19695dc8b7d201d2\": rpc error: code = NotFound desc = could not find container \"e08ab537c3dd40d25131dfb6d6b5843eb734bd97c88efcfd19695dc8b7d201d2\": container with ID starting with e08ab537c3dd40d25131dfb6d6b5843eb734bd97c88efcfd19695dc8b7d201d2 not found: ID does not exist" Oct 03 14:15:17 crc kubenswrapper[4578]: I1003 14:15:17.114427 4578 scope.go:117] "RemoveContainer" containerID="94504235b793cd0adece6abc8d064e56d9bd6eb80afce8c3c77b2fb7fc541a15" Oct 03 14:15:17 crc kubenswrapper[4578]: E1003 14:15:17.114887 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94504235b793cd0adece6abc8d064e56d9bd6eb80afce8c3c77b2fb7fc541a15\": container with ID starting with 94504235b793cd0adece6abc8d064e56d9bd6eb80afce8c3c77b2fb7fc541a15 not found: ID does not exist" containerID="94504235b793cd0adece6abc8d064e56d9bd6eb80afce8c3c77b2fb7fc541a15" Oct 03 14:15:17 crc kubenswrapper[4578]: I1003 14:15:17.114944 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94504235b793cd0adece6abc8d064e56d9bd6eb80afce8c3c77b2fb7fc541a15"} err="failed to get container status \"94504235b793cd0adece6abc8d064e56d9bd6eb80afce8c3c77b2fb7fc541a15\": rpc error: code = NotFound desc = could not find container \"94504235b793cd0adece6abc8d064e56d9bd6eb80afce8c3c77b2fb7fc541a15\": container with ID starting with 94504235b793cd0adece6abc8d064e56d9bd6eb80afce8c3c77b2fb7fc541a15 not found: ID does not exist" Oct 03 14:15:17 crc kubenswrapper[4578]: I1003 14:15:17.348992 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bjs65"] Oct 03 14:15:17 crc kubenswrapper[4578]: I1003 14:15:17.356903 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bjs65"] Oct 03 14:15:18 crc kubenswrapper[4578]: I1003 14:15:18.924091 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="391d359a-b909-4e02-8c37-142072970c22" path="/var/lib/kubelet/pods/391d359a-b909-4e02-8c37-142072970c22/volumes" Oct 03 14:15:22 crc kubenswrapper[4578]: I1003 14:15:22.909360 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:15:22 crc kubenswrapper[4578]: E1003 14:15:22.910177 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:15:34 crc kubenswrapper[4578]: I1003 14:15:34.916028 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:15:34 crc kubenswrapper[4578]: E1003 14:15:34.916864 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:15:49 crc kubenswrapper[4578]: I1003 14:15:49.909827 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:15:49 crc kubenswrapper[4578]: E1003 14:15:49.910527 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:16:02 crc kubenswrapper[4578]: I1003 14:16:02.909502 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:16:02 crc kubenswrapper[4578]: E1003 14:16:02.910327 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:16:14 crc kubenswrapper[4578]: I1003 14:16:14.917341 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:16:14 crc kubenswrapper[4578]: E1003 14:16:14.918131 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:16:27 crc kubenswrapper[4578]: I1003 14:16:27.909111 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:16:27 crc kubenswrapper[4578]: E1003 14:16:27.909893 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:16:41 crc kubenswrapper[4578]: I1003 14:16:41.909327 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:16:41 crc kubenswrapper[4578]: E1003 14:16:41.910110 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:16:55 crc kubenswrapper[4578]: I1003 14:16:55.911792 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:16:55 crc kubenswrapper[4578]: E1003 14:16:55.912726 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:17:04 crc kubenswrapper[4578]: I1003 14:17:04.034667 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x68fn"] Oct 03 14:17:04 crc kubenswrapper[4578]: E1003 14:17:04.035719 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="391d359a-b909-4e02-8c37-142072970c22" containerName="extract-content" Oct 03 14:17:04 crc kubenswrapper[4578]: I1003 14:17:04.035736 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="391d359a-b909-4e02-8c37-142072970c22" containerName="extract-content" Oct 03 14:17:04 crc kubenswrapper[4578]: E1003 14:17:04.035752 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="391d359a-b909-4e02-8c37-142072970c22" containerName="extract-utilities" Oct 03 14:17:04 crc kubenswrapper[4578]: I1003 14:17:04.035759 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="391d359a-b909-4e02-8c37-142072970c22" containerName="extract-utilities" Oct 03 14:17:04 crc kubenswrapper[4578]: E1003 14:17:04.035772 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="391d359a-b909-4e02-8c37-142072970c22" containerName="registry-server" Oct 03 14:17:04 crc kubenswrapper[4578]: I1003 14:17:04.035778 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="391d359a-b909-4e02-8c37-142072970c22" containerName="registry-server" Oct 03 14:17:04 crc kubenswrapper[4578]: I1003 14:17:04.035998 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="391d359a-b909-4e02-8c37-142072970c22" containerName="registry-server" Oct 03 14:17:04 crc kubenswrapper[4578]: I1003 14:17:04.037794 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x68fn" Oct 03 14:17:04 crc kubenswrapper[4578]: I1003 14:17:04.049643 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x68fn"] Oct 03 14:17:04 crc kubenswrapper[4578]: I1003 14:17:04.177463 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eac90d1f-f99f-4572-96ca-d1921afbf7b6-utilities\") pod \"redhat-marketplace-x68fn\" (UID: \"eac90d1f-f99f-4572-96ca-d1921afbf7b6\") " pod="openshift-marketplace/redhat-marketplace-x68fn" Oct 03 14:17:04 crc kubenswrapper[4578]: I1003 14:17:04.177535 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsfpx\" (UniqueName: \"kubernetes.io/projected/eac90d1f-f99f-4572-96ca-d1921afbf7b6-kube-api-access-zsfpx\") pod \"redhat-marketplace-x68fn\" (UID: \"eac90d1f-f99f-4572-96ca-d1921afbf7b6\") " pod="openshift-marketplace/redhat-marketplace-x68fn" Oct 03 14:17:04 crc kubenswrapper[4578]: I1003 14:17:04.177777 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eac90d1f-f99f-4572-96ca-d1921afbf7b6-catalog-content\") pod \"redhat-marketplace-x68fn\" (UID: \"eac90d1f-f99f-4572-96ca-d1921afbf7b6\") " pod="openshift-marketplace/redhat-marketplace-x68fn" Oct 03 14:17:04 crc kubenswrapper[4578]: I1003 14:17:04.279437 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eac90d1f-f99f-4572-96ca-d1921afbf7b6-catalog-content\") pod \"redhat-marketplace-x68fn\" (UID: \"eac90d1f-f99f-4572-96ca-d1921afbf7b6\") " pod="openshift-marketplace/redhat-marketplace-x68fn" Oct 03 14:17:04 crc kubenswrapper[4578]: I1003 14:17:04.279535 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eac90d1f-f99f-4572-96ca-d1921afbf7b6-utilities\") pod \"redhat-marketplace-x68fn\" (UID: \"eac90d1f-f99f-4572-96ca-d1921afbf7b6\") " pod="openshift-marketplace/redhat-marketplace-x68fn" Oct 03 14:17:04 crc kubenswrapper[4578]: I1003 14:17:04.279563 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsfpx\" (UniqueName: \"kubernetes.io/projected/eac90d1f-f99f-4572-96ca-d1921afbf7b6-kube-api-access-zsfpx\") pod \"redhat-marketplace-x68fn\" (UID: \"eac90d1f-f99f-4572-96ca-d1921afbf7b6\") " pod="openshift-marketplace/redhat-marketplace-x68fn" Oct 03 14:17:04 crc kubenswrapper[4578]: I1003 14:17:04.279974 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eac90d1f-f99f-4572-96ca-d1921afbf7b6-catalog-content\") pod \"redhat-marketplace-x68fn\" (UID: \"eac90d1f-f99f-4572-96ca-d1921afbf7b6\") " pod="openshift-marketplace/redhat-marketplace-x68fn" Oct 03 14:17:04 crc kubenswrapper[4578]: I1003 14:17:04.280225 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eac90d1f-f99f-4572-96ca-d1921afbf7b6-utilities\") pod \"redhat-marketplace-x68fn\" (UID: \"eac90d1f-f99f-4572-96ca-d1921afbf7b6\") " pod="openshift-marketplace/redhat-marketplace-x68fn" Oct 03 14:17:04 crc kubenswrapper[4578]: I1003 14:17:04.315872 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsfpx\" (UniqueName: \"kubernetes.io/projected/eac90d1f-f99f-4572-96ca-d1921afbf7b6-kube-api-access-zsfpx\") pod \"redhat-marketplace-x68fn\" (UID: \"eac90d1f-f99f-4572-96ca-d1921afbf7b6\") " pod="openshift-marketplace/redhat-marketplace-x68fn" Oct 03 14:17:04 crc kubenswrapper[4578]: I1003 14:17:04.367694 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x68fn" Oct 03 14:17:04 crc kubenswrapper[4578]: I1003 14:17:04.856280 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x68fn"] Oct 03 14:17:05 crc kubenswrapper[4578]: I1003 14:17:04.999972 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x68fn" event={"ID":"eac90d1f-f99f-4572-96ca-d1921afbf7b6","Type":"ContainerStarted","Data":"448bc2bb0216b579ec5542c166dd1683c1e57bb0a4e6846e8715df2505d60a97"} Oct 03 14:17:06 crc kubenswrapper[4578]: I1003 14:17:06.012488 4578 generic.go:334] "Generic (PLEG): container finished" podID="eac90d1f-f99f-4572-96ca-d1921afbf7b6" containerID="a226ce0ff3928d2403efe18ad819362ac5646ba1397fbbbfbbac25d0a003c9ad" exitCode=0 Oct 03 14:17:06 crc kubenswrapper[4578]: I1003 14:17:06.012658 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x68fn" event={"ID":"eac90d1f-f99f-4572-96ca-d1921afbf7b6","Type":"ContainerDied","Data":"a226ce0ff3928d2403efe18ad819362ac5646ba1397fbbbfbbac25d0a003c9ad"} Oct 03 14:17:06 crc kubenswrapper[4578]: I1003 14:17:06.020019 4578 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 03 14:17:06 crc kubenswrapper[4578]: I1003 14:17:06.909511 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:17:06 crc kubenswrapper[4578]: E1003 14:17:06.910170 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:17:08 crc kubenswrapper[4578]: I1003 14:17:08.030782 4578 generic.go:334] "Generic (PLEG): container finished" podID="eac90d1f-f99f-4572-96ca-d1921afbf7b6" containerID="a06376bccef80d4ef16699cf34152e3a53980aae0275e1f30585eaf072106b63" exitCode=0 Oct 03 14:17:08 crc kubenswrapper[4578]: I1003 14:17:08.031017 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x68fn" event={"ID":"eac90d1f-f99f-4572-96ca-d1921afbf7b6","Type":"ContainerDied","Data":"a06376bccef80d4ef16699cf34152e3a53980aae0275e1f30585eaf072106b63"} Oct 03 14:17:10 crc kubenswrapper[4578]: I1003 14:17:10.052154 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x68fn" event={"ID":"eac90d1f-f99f-4572-96ca-d1921afbf7b6","Type":"ContainerStarted","Data":"9c83a7d30ad1054fab8bd2ee250c7fb0de7ae4b9111a154195cc50f3fe7e22c7"} Oct 03 14:17:10 crc kubenswrapper[4578]: I1003 14:17:10.072167 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x68fn" podStartSLOduration=3.132139633 podStartE2EDuration="6.072145008s" podCreationTimestamp="2025-10-03 14:17:04 +0000 UTC" firstStartedPulling="2025-10-03 14:17:06.019796992 +0000 UTC m=+5161.818269176" lastFinishedPulling="2025-10-03 14:17:08.959802367 +0000 UTC m=+5164.758274551" observedRunningTime="2025-10-03 14:17:10.069911538 +0000 UTC m=+5165.868383742" watchObservedRunningTime="2025-10-03 14:17:10.072145008 +0000 UTC m=+5165.870617192" Oct 03 14:17:14 crc kubenswrapper[4578]: I1003 14:17:14.242152 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mcgpg/must-gather-2rhkb"] Oct 03 14:17:14 crc kubenswrapper[4578]: I1003 14:17:14.244545 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mcgpg/must-gather-2rhkb" Oct 03 14:17:14 crc kubenswrapper[4578]: I1003 14:17:14.255694 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mcgpg"/"openshift-service-ca.crt" Oct 03 14:17:14 crc kubenswrapper[4578]: I1003 14:17:14.256032 4578 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mcgpg"/"kube-root-ca.crt" Oct 03 14:17:14 crc kubenswrapper[4578]: I1003 14:17:14.256209 4578 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mcgpg"/"default-dockercfg-lb9rb" Oct 03 14:17:14 crc kubenswrapper[4578]: I1003 14:17:14.276750 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mcgpg/must-gather-2rhkb"] Oct 03 14:17:14 crc kubenswrapper[4578]: I1003 14:17:14.368280 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x68fn" Oct 03 14:17:14 crc kubenswrapper[4578]: I1003 14:17:14.369059 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x68fn" Oct 03 14:17:14 crc kubenswrapper[4578]: I1003 14:17:14.373467 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjp46\" (UniqueName: \"kubernetes.io/projected/c7e530e9-622a-4d8a-b2aa-88d8af716b77-kube-api-access-qjp46\") pod \"must-gather-2rhkb\" (UID: \"c7e530e9-622a-4d8a-b2aa-88d8af716b77\") " pod="openshift-must-gather-mcgpg/must-gather-2rhkb" Oct 03 14:17:14 crc kubenswrapper[4578]: I1003 14:17:14.373711 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c7e530e9-622a-4d8a-b2aa-88d8af716b77-must-gather-output\") pod \"must-gather-2rhkb\" (UID: \"c7e530e9-622a-4d8a-b2aa-88d8af716b77\") " pod="openshift-must-gather-mcgpg/must-gather-2rhkb" Oct 03 14:17:14 crc kubenswrapper[4578]: I1003 14:17:14.439959 4578 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x68fn" Oct 03 14:17:14 crc kubenswrapper[4578]: I1003 14:17:14.475315 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c7e530e9-622a-4d8a-b2aa-88d8af716b77-must-gather-output\") pod \"must-gather-2rhkb\" (UID: \"c7e530e9-622a-4d8a-b2aa-88d8af716b77\") " pod="openshift-must-gather-mcgpg/must-gather-2rhkb" Oct 03 14:17:14 crc kubenswrapper[4578]: I1003 14:17:14.475413 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjp46\" (UniqueName: \"kubernetes.io/projected/c7e530e9-622a-4d8a-b2aa-88d8af716b77-kube-api-access-qjp46\") pod \"must-gather-2rhkb\" (UID: \"c7e530e9-622a-4d8a-b2aa-88d8af716b77\") " pod="openshift-must-gather-mcgpg/must-gather-2rhkb" Oct 03 14:17:14 crc kubenswrapper[4578]: I1003 14:17:14.476203 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c7e530e9-622a-4d8a-b2aa-88d8af716b77-must-gather-output\") pod \"must-gather-2rhkb\" (UID: \"c7e530e9-622a-4d8a-b2aa-88d8af716b77\") " pod="openshift-must-gather-mcgpg/must-gather-2rhkb" Oct 03 14:17:14 crc kubenswrapper[4578]: I1003 14:17:14.506454 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjp46\" (UniqueName: \"kubernetes.io/projected/c7e530e9-622a-4d8a-b2aa-88d8af716b77-kube-api-access-qjp46\") pod \"must-gather-2rhkb\" (UID: \"c7e530e9-622a-4d8a-b2aa-88d8af716b77\") " pod="openshift-must-gather-mcgpg/must-gather-2rhkb" Oct 03 14:17:14 crc kubenswrapper[4578]: I1003 14:17:14.593229 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mcgpg/must-gather-2rhkb" Oct 03 14:17:15 crc kubenswrapper[4578]: I1003 14:17:15.187237 4578 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mcgpg/must-gather-2rhkb"] Oct 03 14:17:15 crc kubenswrapper[4578]: I1003 14:17:15.327172 4578 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x68fn" Oct 03 14:17:15 crc kubenswrapper[4578]: I1003 14:17:15.388758 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x68fn"] Oct 03 14:17:16 crc kubenswrapper[4578]: I1003 14:17:16.110369 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mcgpg/must-gather-2rhkb" event={"ID":"c7e530e9-622a-4d8a-b2aa-88d8af716b77","Type":"ContainerStarted","Data":"b563ef6487ce9c9cce24cbf7b87f80f403535bb03ebb7a7e3103ff093a437868"} Oct 03 14:17:17 crc kubenswrapper[4578]: I1003 14:17:17.118353 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x68fn" podUID="eac90d1f-f99f-4572-96ca-d1921afbf7b6" containerName="registry-server" containerID="cri-o://9c83a7d30ad1054fab8bd2ee250c7fb0de7ae4b9111a154195cc50f3fe7e22c7" gracePeriod=2 Oct 03 14:17:18 crc kubenswrapper[4578]: I1003 14:17:18.147236 4578 generic.go:334] "Generic (PLEG): container finished" podID="eac90d1f-f99f-4572-96ca-d1921afbf7b6" containerID="9c83a7d30ad1054fab8bd2ee250c7fb0de7ae4b9111a154195cc50f3fe7e22c7" exitCode=0 Oct 03 14:17:18 crc kubenswrapper[4578]: I1003 14:17:18.147312 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x68fn" event={"ID":"eac90d1f-f99f-4572-96ca-d1921afbf7b6","Type":"ContainerDied","Data":"9c83a7d30ad1054fab8bd2ee250c7fb0de7ae4b9111a154195cc50f3fe7e22c7"} Oct 03 14:17:18 crc kubenswrapper[4578]: I1003 14:17:18.361250 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x68fn" Oct 03 14:17:18 crc kubenswrapper[4578]: I1003 14:17:18.486427 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eac90d1f-f99f-4572-96ca-d1921afbf7b6-catalog-content\") pod \"eac90d1f-f99f-4572-96ca-d1921afbf7b6\" (UID: \"eac90d1f-f99f-4572-96ca-d1921afbf7b6\") " Oct 03 14:17:18 crc kubenswrapper[4578]: I1003 14:17:18.486525 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eac90d1f-f99f-4572-96ca-d1921afbf7b6-utilities\") pod \"eac90d1f-f99f-4572-96ca-d1921afbf7b6\" (UID: \"eac90d1f-f99f-4572-96ca-d1921afbf7b6\") " Oct 03 14:17:18 crc kubenswrapper[4578]: I1003 14:17:18.486609 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsfpx\" (UniqueName: \"kubernetes.io/projected/eac90d1f-f99f-4572-96ca-d1921afbf7b6-kube-api-access-zsfpx\") pod \"eac90d1f-f99f-4572-96ca-d1921afbf7b6\" (UID: \"eac90d1f-f99f-4572-96ca-d1921afbf7b6\") " Oct 03 14:17:18 crc kubenswrapper[4578]: I1003 14:17:18.487475 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eac90d1f-f99f-4572-96ca-d1921afbf7b6-utilities" (OuterVolumeSpecName: "utilities") pod "eac90d1f-f99f-4572-96ca-d1921afbf7b6" (UID: "eac90d1f-f99f-4572-96ca-d1921afbf7b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:17:18 crc kubenswrapper[4578]: I1003 14:17:18.499096 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eac90d1f-f99f-4572-96ca-d1921afbf7b6-kube-api-access-zsfpx" (OuterVolumeSpecName: "kube-api-access-zsfpx") pod "eac90d1f-f99f-4572-96ca-d1921afbf7b6" (UID: "eac90d1f-f99f-4572-96ca-d1921afbf7b6"). InnerVolumeSpecName "kube-api-access-zsfpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:17:18 crc kubenswrapper[4578]: I1003 14:17:18.505477 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eac90d1f-f99f-4572-96ca-d1921afbf7b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eac90d1f-f99f-4572-96ca-d1921afbf7b6" (UID: "eac90d1f-f99f-4572-96ca-d1921afbf7b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:17:18 crc kubenswrapper[4578]: I1003 14:17:18.589068 4578 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eac90d1f-f99f-4572-96ca-d1921afbf7b6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 03 14:17:18 crc kubenswrapper[4578]: I1003 14:17:18.589104 4578 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eac90d1f-f99f-4572-96ca-d1921afbf7b6-utilities\") on node \"crc\" DevicePath \"\"" Oct 03 14:17:18 crc kubenswrapper[4578]: I1003 14:17:18.589113 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsfpx\" (UniqueName: \"kubernetes.io/projected/eac90d1f-f99f-4572-96ca-d1921afbf7b6-kube-api-access-zsfpx\") on node \"crc\" DevicePath \"\"" Oct 03 14:17:19 crc kubenswrapper[4578]: I1003 14:17:19.161687 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x68fn" event={"ID":"eac90d1f-f99f-4572-96ca-d1921afbf7b6","Type":"ContainerDied","Data":"448bc2bb0216b579ec5542c166dd1683c1e57bb0a4e6846e8715df2505d60a97"} Oct 03 14:17:19 crc kubenswrapper[4578]: I1003 14:17:19.162042 4578 scope.go:117] "RemoveContainer" containerID="9c83a7d30ad1054fab8bd2ee250c7fb0de7ae4b9111a154195cc50f3fe7e22c7" Oct 03 14:17:19 crc kubenswrapper[4578]: I1003 14:17:19.161783 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x68fn" Oct 03 14:17:19 crc kubenswrapper[4578]: I1003 14:17:19.193819 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x68fn"] Oct 03 14:17:19 crc kubenswrapper[4578]: I1003 14:17:19.203756 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x68fn"] Oct 03 14:17:20 crc kubenswrapper[4578]: I1003 14:17:20.909508 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:17:20 crc kubenswrapper[4578]: E1003 14:17:20.910181 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:17:20 crc kubenswrapper[4578]: I1003 14:17:20.930191 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eac90d1f-f99f-4572-96ca-d1921afbf7b6" path="/var/lib/kubelet/pods/eac90d1f-f99f-4572-96ca-d1921afbf7b6/volumes" Oct 03 14:17:21 crc kubenswrapper[4578]: I1003 14:17:21.281089 4578 scope.go:117] "RemoveContainer" containerID="a06376bccef80d4ef16699cf34152e3a53980aae0275e1f30585eaf072106b63" Oct 03 14:17:21 crc kubenswrapper[4578]: I1003 14:17:21.302550 4578 scope.go:117] "RemoveContainer" containerID="a226ce0ff3928d2403efe18ad819362ac5646ba1397fbbbfbbac25d0a003c9ad" Oct 03 14:17:26 crc kubenswrapper[4578]: I1003 14:17:26.226538 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mcgpg/must-gather-2rhkb" event={"ID":"c7e530e9-622a-4d8a-b2aa-88d8af716b77","Type":"ContainerStarted","Data":"0b171dd5a470b945fed5a6e01b36368ba15e66a41647104cb565ecccbadf8fa3"} Oct 03 14:17:27 crc kubenswrapper[4578]: I1003 14:17:27.234898 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mcgpg/must-gather-2rhkb" event={"ID":"c7e530e9-622a-4d8a-b2aa-88d8af716b77","Type":"ContainerStarted","Data":"f8bcbc6e4e9f8d5b48f159903251abf4f22e825103a8665e3592c7b5992b1ca7"} Oct 03 14:17:27 crc kubenswrapper[4578]: I1003 14:17:27.262661 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mcgpg/must-gather-2rhkb" podStartSLOduration=3.808253976 podStartE2EDuration="13.262642843s" podCreationTimestamp="2025-10-03 14:17:14 +0000 UTC" firstStartedPulling="2025-10-03 14:17:15.202663949 +0000 UTC m=+5171.001136133" lastFinishedPulling="2025-10-03 14:17:24.657052816 +0000 UTC m=+5180.455525000" observedRunningTime="2025-10-03 14:17:27.258001608 +0000 UTC m=+5183.056473792" watchObservedRunningTime="2025-10-03 14:17:27.262642843 +0000 UTC m=+5183.061115027" Oct 03 14:17:33 crc kubenswrapper[4578]: I1003 14:17:33.909378 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:17:33 crc kubenswrapper[4578]: E1003 14:17:33.909978 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:17:34 crc kubenswrapper[4578]: I1003 14:17:34.768619 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mcgpg/crc-debug-jjz6m"] Oct 03 14:17:34 crc kubenswrapper[4578]: E1003 14:17:34.769442 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eac90d1f-f99f-4572-96ca-d1921afbf7b6" containerName="extract-utilities" Oct 03 14:17:34 crc kubenswrapper[4578]: I1003 14:17:34.769459 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="eac90d1f-f99f-4572-96ca-d1921afbf7b6" containerName="extract-utilities" Oct 03 14:17:34 crc kubenswrapper[4578]: E1003 14:17:34.769484 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eac90d1f-f99f-4572-96ca-d1921afbf7b6" containerName="extract-content" Oct 03 14:17:34 crc kubenswrapper[4578]: I1003 14:17:34.769490 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="eac90d1f-f99f-4572-96ca-d1921afbf7b6" containerName="extract-content" Oct 03 14:17:34 crc kubenswrapper[4578]: E1003 14:17:34.769515 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eac90d1f-f99f-4572-96ca-d1921afbf7b6" containerName="registry-server" Oct 03 14:17:34 crc kubenswrapper[4578]: I1003 14:17:34.769522 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="eac90d1f-f99f-4572-96ca-d1921afbf7b6" containerName="registry-server" Oct 03 14:17:34 crc kubenswrapper[4578]: I1003 14:17:34.769770 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="eac90d1f-f99f-4572-96ca-d1921afbf7b6" containerName="registry-server" Oct 03 14:17:34 crc kubenswrapper[4578]: I1003 14:17:34.770441 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mcgpg/crc-debug-jjz6m" Oct 03 14:17:34 crc kubenswrapper[4578]: I1003 14:17:34.915582 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxqxh\" (UniqueName: \"kubernetes.io/projected/0d5b463e-66dd-4d15-8713-6a762ef6fdc4-kube-api-access-qxqxh\") pod \"crc-debug-jjz6m\" (UID: \"0d5b463e-66dd-4d15-8713-6a762ef6fdc4\") " pod="openshift-must-gather-mcgpg/crc-debug-jjz6m" Oct 03 14:17:34 crc kubenswrapper[4578]: I1003 14:17:34.915661 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d5b463e-66dd-4d15-8713-6a762ef6fdc4-host\") pod \"crc-debug-jjz6m\" (UID: \"0d5b463e-66dd-4d15-8713-6a762ef6fdc4\") " pod="openshift-must-gather-mcgpg/crc-debug-jjz6m" Oct 03 14:17:35 crc kubenswrapper[4578]: I1003 14:17:35.017142 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxqxh\" (UniqueName: \"kubernetes.io/projected/0d5b463e-66dd-4d15-8713-6a762ef6fdc4-kube-api-access-qxqxh\") pod \"crc-debug-jjz6m\" (UID: \"0d5b463e-66dd-4d15-8713-6a762ef6fdc4\") " pod="openshift-must-gather-mcgpg/crc-debug-jjz6m" Oct 03 14:17:35 crc kubenswrapper[4578]: I1003 14:17:35.017229 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d5b463e-66dd-4d15-8713-6a762ef6fdc4-host\") pod \"crc-debug-jjz6m\" (UID: \"0d5b463e-66dd-4d15-8713-6a762ef6fdc4\") " pod="openshift-must-gather-mcgpg/crc-debug-jjz6m" Oct 03 14:17:35 crc kubenswrapper[4578]: I1003 14:17:35.017942 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d5b463e-66dd-4d15-8713-6a762ef6fdc4-host\") pod \"crc-debug-jjz6m\" (UID: \"0d5b463e-66dd-4d15-8713-6a762ef6fdc4\") " pod="openshift-must-gather-mcgpg/crc-debug-jjz6m" Oct 03 14:17:35 crc kubenswrapper[4578]: I1003 14:17:35.040731 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxqxh\" (UniqueName: \"kubernetes.io/projected/0d5b463e-66dd-4d15-8713-6a762ef6fdc4-kube-api-access-qxqxh\") pod \"crc-debug-jjz6m\" (UID: \"0d5b463e-66dd-4d15-8713-6a762ef6fdc4\") " pod="openshift-must-gather-mcgpg/crc-debug-jjz6m" Oct 03 14:17:35 crc kubenswrapper[4578]: I1003 14:17:35.096344 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mcgpg/crc-debug-jjz6m" Oct 03 14:17:35 crc kubenswrapper[4578]: I1003 14:17:35.318837 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mcgpg/crc-debug-jjz6m" event={"ID":"0d5b463e-66dd-4d15-8713-6a762ef6fdc4","Type":"ContainerStarted","Data":"f5cbec1ecbc5b5db367b5a5f3ca840aca8d2f4889ad05634c45c604ec9b16ba6"} Oct 03 14:17:46 crc kubenswrapper[4578]: I1003 14:17:46.909354 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:17:46 crc kubenswrapper[4578]: E1003 14:17:46.910171 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:17:47 crc kubenswrapper[4578]: I1003 14:17:47.431263 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mcgpg/crc-debug-jjz6m" event={"ID":"0d5b463e-66dd-4d15-8713-6a762ef6fdc4","Type":"ContainerStarted","Data":"4269a4450af0f08820987b8080dd0e13e068a1c7afa451e8a8bd73bb1476c239"} Oct 03 14:17:47 crc kubenswrapper[4578]: I1003 14:17:47.444746 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mcgpg/crc-debug-jjz6m" podStartSLOduration=1.891404454 podStartE2EDuration="13.444730739s" podCreationTimestamp="2025-10-03 14:17:34 +0000 UTC" firstStartedPulling="2025-10-03 14:17:35.13816514 +0000 UTC m=+5190.936637324" lastFinishedPulling="2025-10-03 14:17:46.691491425 +0000 UTC m=+5202.489963609" observedRunningTime="2025-10-03 14:17:47.44317745 +0000 UTC m=+5203.241649634" watchObservedRunningTime="2025-10-03 14:17:47.444730739 +0000 UTC m=+5203.243202923" Oct 03 14:17:57 crc kubenswrapper[4578]: I1003 14:17:57.910000 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:17:57 crc kubenswrapper[4578]: E1003 14:17:57.910887 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:18:08 crc kubenswrapper[4578]: I1003 14:18:08.909707 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:18:08 crc kubenswrapper[4578]: E1003 14:18:08.910495 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:18:19 crc kubenswrapper[4578]: I1003 14:18:19.909859 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:18:19 crc kubenswrapper[4578]: E1003 14:18:19.910603 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:18:31 crc kubenswrapper[4578]: I1003 14:18:31.909580 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:18:31 crc kubenswrapper[4578]: E1003 14:18:31.910384 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:18:43 crc kubenswrapper[4578]: I1003 14:18:43.909755 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:18:45 crc kubenswrapper[4578]: I1003 14:18:45.054821 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"99c6af77c4906d954ca12c7460c666fc7a526bfb23a532ab77b5add6bbd7dead"} Oct 03 14:18:57 crc kubenswrapper[4578]: I1003 14:18:57.775248 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-f87d658c4-wf95n_8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922/barbican-api/0.log" Oct 03 14:18:57 crc kubenswrapper[4578]: I1003 14:18:57.881077 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-f87d658c4-wf95n_8d04d0a3-7bb8-46f8-bcb8-a3ceb49ba922/barbican-api-log/0.log" Oct 03 14:18:58 crc kubenswrapper[4578]: I1003 14:18:58.028283 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-786f64db8b-9n26w_3bb09a8d-f51b-494d-8a75-8690e842a916/barbican-keystone-listener/0.log" Oct 03 14:18:58 crc kubenswrapper[4578]: I1003 14:18:58.130885 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-786f64db8b-9n26w_3bb09a8d-f51b-494d-8a75-8690e842a916/barbican-keystone-listener-log/0.log" Oct 03 14:18:58 crc kubenswrapper[4578]: I1003 14:18:58.233543 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-d9f65c5d7-qf4l9_c187292d-a86d-455d-9c52-dabc5c3b4631/barbican-worker/0.log" Oct 03 14:18:58 crc kubenswrapper[4578]: I1003 14:18:58.355421 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-d9f65c5d7-qf4l9_c187292d-a86d-455d-9c52-dabc5c3b4631/barbican-worker-log/0.log" Oct 03 14:18:58 crc kubenswrapper[4578]: I1003 14:18:58.581979 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-46ct8_0095f720-7bf0-40fe-8f0c-e70fa5fc65d4/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 14:18:58 crc kubenswrapper[4578]: I1003 14:18:58.729548 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a54925d7-6f48-4ce4-8062-c81a33abb815/ceilometer-central-agent/0.log" Oct 03 14:18:58 crc kubenswrapper[4578]: I1003 14:18:58.790622 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a54925d7-6f48-4ce4-8062-c81a33abb815/ceilometer-notification-agent/0.log" Oct 03 14:18:58 crc kubenswrapper[4578]: I1003 14:18:58.830342 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a54925d7-6f48-4ce4-8062-c81a33abb815/proxy-httpd/0.log" Oct 03 14:18:58 crc kubenswrapper[4578]: I1003 14:18:58.953305 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a54925d7-6f48-4ce4-8062-c81a33abb815/sg-core/0.log" Oct 03 14:18:59 crc kubenswrapper[4578]: I1003 14:18:59.173543 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_eb507b70-6b06-42c0-aec3-e5343f719efe/cinder-api/0.log" Oct 03 14:18:59 crc kubenswrapper[4578]: I1003 14:18:59.221523 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_eb507b70-6b06-42c0-aec3-e5343f719efe/cinder-api-log/0.log" Oct 03 14:18:59 crc kubenswrapper[4578]: I1003 14:18:59.460509 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c380ba54-4a4d-4e1f-8f75-4628c1f33d7c/cinder-scheduler/0.log" Oct 03 14:18:59 crc kubenswrapper[4578]: I1003 14:18:59.472523 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c380ba54-4a4d-4e1f-8f75-4628c1f33d7c/probe/0.log" Oct 03 14:18:59 crc kubenswrapper[4578]: I1003 14:18:59.924784 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-nm2pv_70bd4887-f262-43e1-bd62-2a4f1727c67b/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 14:19:00 crc kubenswrapper[4578]: I1003 14:19:00.047145 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-7fctt_53e7ff32-dd82-4be9-b309-7cd2b0130d32/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 14:19:00 crc kubenswrapper[4578]: I1003 14:19:00.185290 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-bl2fg_6b63f0ce-edc6-4405-8ec4-7d1b9c08dd1e/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 14:19:00 crc kubenswrapper[4578]: I1003 14:19:00.400030 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-82gmw_76bba854-23ba-4bf1-b4aa-db525ff8c5d8/init/0.log" Oct 03 14:19:00 crc kubenswrapper[4578]: I1003 14:19:00.752124 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-82gmw_76bba854-23ba-4bf1-b4aa-db525ff8c5d8/dnsmasq-dns/0.log" Oct 03 14:19:00 crc kubenswrapper[4578]: I1003 14:19:00.759914 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-79dc84bdb7-82gmw_76bba854-23ba-4bf1-b4aa-db525ff8c5d8/init/0.log" Oct 03 14:19:01 crc kubenswrapper[4578]: I1003 14:19:01.407140 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_3d110456-18c3-4369-9ca7-0efcf6a02d8b/glance-httpd/0.log" Oct 03 14:19:01 crc kubenswrapper[4578]: I1003 14:19:01.508819 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-x6mpg_79b337ea-fd3a-49a6-9f82-67e1a86b29b0/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 14:19:01 crc kubenswrapper[4578]: I1003 14:19:01.702611 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_3d110456-18c3-4369-9ca7-0efcf6a02d8b/glance-log/0.log" Oct 03 14:19:01 crc kubenswrapper[4578]: I1003 14:19:01.819857 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_92636290-70b5-44e9-94f2-988b490ae46c/glance-httpd/0.log" Oct 03 14:19:01 crc kubenswrapper[4578]: I1003 14:19:01.942264 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_92636290-70b5-44e9-94f2-988b490ae46c/glance-log/0.log" Oct 03 14:19:02 crc kubenswrapper[4578]: I1003 14:19:02.134509 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-68684899bd-2cwll_d49d623e-2cd6-4786-b1da-34b78b89a134/horizon/1.log" Oct 03 14:19:02 crc kubenswrapper[4578]: I1003 14:19:02.365080 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-68684899bd-2cwll_d49d623e-2cd6-4786-b1da-34b78b89a134/horizon/0.log" Oct 03 14:19:02 crc kubenswrapper[4578]: I1003 14:19:02.795821 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-trqwp_76db17db-5757-4c06-94d4-537c645f54e5/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 14:19:03 crc kubenswrapper[4578]: I1003 14:19:03.045305 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-68684899bd-2cwll_d49d623e-2cd6-4786-b1da-34b78b89a134/horizon-log/0.log" Oct 03 14:19:03 crc kubenswrapper[4578]: I1003 14:19:03.107444 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-jnfmm_ec1fc355-06aa-42cb-9b25-3b7ef42e0ba4/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 14:19:03 crc kubenswrapper[4578]: I1003 14:19:03.299366 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-64b9f57cc5-jrm9t_f70c10e6-f3e1-41e0-a30f-267ca3a05c82/keystone-api/0.log" Oct 03 14:19:03 crc kubenswrapper[4578]: I1003 14:19:03.420292 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29325001-hhn29_7d6d4828-38c5-41cf-93a4-a25d8a9810a2/keystone-cron/0.log" Oct 03 14:19:03 crc kubenswrapper[4578]: I1003 14:19:03.629997 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_95fd518d-9176-4f37-ba66-e636d9794533/kube-state-metrics/0.log" Oct 03 14:19:03 crc kubenswrapper[4578]: I1003 14:19:03.986898 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-d7d5c9485-l2ww5_2711a2e9-8d7a-4942-b762-f211d519e0f4/neutron-api/0.log" Oct 03 14:19:04 crc kubenswrapper[4578]: I1003 14:19:04.340419 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-d7d5c9485-l2ww5_2711a2e9-8d7a-4942-b762-f211d519e0f4/neutron-httpd/0.log" Oct 03 14:19:04 crc kubenswrapper[4578]: I1003 14:19:04.846755 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_25b861ed-9dfe-4494-96a0-03fa994f1775/nova-api-log/0.log" Oct 03 14:19:04 crc kubenswrapper[4578]: I1003 14:19:04.913928 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_25b861ed-9dfe-4494-96a0-03fa994f1775/nova-api-api/0.log" Oct 03 14:19:05 crc kubenswrapper[4578]: I1003 14:19:05.515171 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_fdbce8e5-76ed-4e0c-9649-5c8c783f0986/nova-cell1-conductor-conductor/0.log" Oct 03 14:19:05 crc kubenswrapper[4578]: I1003 14:19:05.523869 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_cf61a63d-477b-4e6d-9946-fd268045d303/nova-cell0-conductor-conductor/0.log" Oct 03 14:19:06 crc kubenswrapper[4578]: I1003 14:19:06.040408 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_624416e7-f2e2-462e-a0f1-e06ce21514f3/nova-cell1-novncproxy-novncproxy/0.log" Oct 03 14:19:06 crc kubenswrapper[4578]: I1003 14:19:06.263933 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_89f4ced5-a0c4-4b41-9c87-64377d0fc4bd/nova-metadata-log/0.log" Oct 03 14:19:06 crc kubenswrapper[4578]: I1003 14:19:06.700760 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_5b3f582b-6953-464d-902b-f008def7a0fe/nova-scheduler-scheduler/0.log" Oct 03 14:19:06 crc kubenswrapper[4578]: I1003 14:19:06.949817 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_32caee81-c7c2-45ee-bb87-2e6003d6115d/mysql-bootstrap/0.log" Oct 03 14:19:07 crc kubenswrapper[4578]: I1003 14:19:07.251455 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_32caee81-c7c2-45ee-bb87-2e6003d6115d/mysql-bootstrap/0.log" Oct 03 14:19:07 crc kubenswrapper[4578]: I1003 14:19:07.276980 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_32caee81-c7c2-45ee-bb87-2e6003d6115d/galera/0.log" Oct 03 14:19:07 crc kubenswrapper[4578]: I1003 14:19:07.635728 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_62242ee6-b3ff-4771-a989-4d7bd1d243d3/mysql-bootstrap/0.log" Oct 03 14:19:07 crc kubenswrapper[4578]: I1003 14:19:07.931922 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_62242ee6-b3ff-4771-a989-4d7bd1d243d3/mysql-bootstrap/0.log" Oct 03 14:19:07 crc kubenswrapper[4578]: I1003 14:19:07.973049 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_62242ee6-b3ff-4771-a989-4d7bd1d243d3/galera/0.log" Oct 03 14:19:08 crc kubenswrapper[4578]: I1003 14:19:08.295687 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_f409b2c2-1191-4048-b0bc-e5c449425aa8/openstackclient/0.log" Oct 03 14:19:08 crc kubenswrapper[4578]: I1003 14:19:08.469567 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_89f4ced5-a0c4-4b41-9c87-64377d0fc4bd/nova-metadata-metadata/0.log" Oct 03 14:19:08 crc kubenswrapper[4578]: I1003 14:19:08.553258 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-9jl4l_daabdc14-9eaa-478d-8e0f-3ab92c0568ce/ovn-controller/0.log" Oct 03 14:19:08 crc kubenswrapper[4578]: I1003 14:19:08.905083 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-bl5dj_d2ac114a-65da-41e3-9067-f076c7fbb9c6/openstack-network-exporter/0.log" Oct 03 14:19:09 crc kubenswrapper[4578]: I1003 14:19:09.353543 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2sqbb_14bc2fd3-f370-48d0-8711-168b0258f21f/ovsdb-server-init/0.log" Oct 03 14:19:09 crc kubenswrapper[4578]: I1003 14:19:09.589930 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2sqbb_14bc2fd3-f370-48d0-8711-168b0258f21f/ovsdb-server/0.log" Oct 03 14:19:09 crc kubenswrapper[4578]: I1003 14:19:09.632850 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2sqbb_14bc2fd3-f370-48d0-8711-168b0258f21f/ovsdb-server-init/0.log" Oct 03 14:19:09 crc kubenswrapper[4578]: I1003 14:19:09.690514 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2sqbb_14bc2fd3-f370-48d0-8711-168b0258f21f/ovs-vswitchd/0.log" Oct 03 14:19:09 crc kubenswrapper[4578]: I1003 14:19:09.935693 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-7p4mv_8856883f-4c07-432c-aadf-399e845876cf/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 14:19:10 crc kubenswrapper[4578]: I1003 14:19:10.175206 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-7ztl6_e6ee1a77-57be-48c1-a61f-50719bfcbd29/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 14:19:10 crc kubenswrapper[4578]: I1003 14:19:10.275732 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-dbcwl_3a6aa5b7-9383-4465-b597-f7330e091dcc/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 14:19:10 crc kubenswrapper[4578]: I1003 14:19:10.627645 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-dqxfw_084860df-b3eb-43f8-aa12-2b7feea78f6a/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 14:19:10 crc kubenswrapper[4578]: I1003 14:19:10.936091 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-k586j_4df021ef-2d0b-4450-9c37-bd35443a3b1a/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 14:19:11 crc kubenswrapper[4578]: I1003 14:19:11.033813 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-tnz6b_2ee86abf-8757-4c4e-8841-37d978cff99a/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 14:19:11 crc kubenswrapper[4578]: I1003 14:19:11.418857 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-wn5jh_f394e983-821e-4983-a093-711f2d6e0a23/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 14:19:11 crc kubenswrapper[4578]: I1003 14:19:11.491930 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_54265d19-6511-40ac-9393-4cf7827def68/openstack-network-exporter/0.log" Oct 03 14:19:11 crc kubenswrapper[4578]: I1003 14:19:11.642026 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_54265d19-6511-40ac-9393-4cf7827def68/ovn-northd/0.log" Oct 03 14:19:11 crc kubenswrapper[4578]: I1003 14:19:11.841309 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_df1fc342-37e6-4757-b1cf-81a10245997b/openstack-network-exporter/0.log" Oct 03 14:19:11 crc kubenswrapper[4578]: I1003 14:19:11.965044 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_df1fc342-37e6-4757-b1cf-81a10245997b/ovsdbserver-nb/0.log" Oct 03 14:19:12 crc kubenswrapper[4578]: I1003 14:19:12.243232 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_efb87daa-8d73-444a-94ed-9e44006bec59/openstack-network-exporter/0.log" Oct 03 14:19:12 crc kubenswrapper[4578]: I1003 14:19:12.282408 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_efb87daa-8d73-444a-94ed-9e44006bec59/ovsdbserver-sb/0.log" Oct 03 14:19:12 crc kubenswrapper[4578]: I1003 14:19:12.524580 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-66cbcb46d8-9jksg_5e89457e-7c3d-4f85-a14d-ca47f5bdca07/placement-api/0.log" Oct 03 14:19:12 crc kubenswrapper[4578]: I1003 14:19:12.988901 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-66cbcb46d8-9jksg_5e89457e-7c3d-4f85-a14d-ca47f5bdca07/placement-log/0.log" Oct 03 14:19:13 crc kubenswrapper[4578]: I1003 14:19:13.299566 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_56e6269a-6d4b-4e78-9fef-7224277e227a/setup-container/0.log" Oct 03 14:19:13 crc kubenswrapper[4578]: I1003 14:19:13.477000 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_56e6269a-6d4b-4e78-9fef-7224277e227a/setup-container/0.log" Oct 03 14:19:13 crc kubenswrapper[4578]: I1003 14:19:13.522456 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_56e6269a-6d4b-4e78-9fef-7224277e227a/rabbitmq/0.log" Oct 03 14:19:13 crc kubenswrapper[4578]: I1003 14:19:13.784738 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_dc895ffd-a4ba-497b-8600-c8491f007547/setup-container/0.log" Oct 03 14:19:14 crc kubenswrapper[4578]: I1003 14:19:14.074458 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_dc895ffd-a4ba-497b-8600-c8491f007547/setup-container/0.log" Oct 03 14:19:14 crc kubenswrapper[4578]: I1003 14:19:14.196759 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_dc895ffd-a4ba-497b-8600-c8491f007547/rabbitmq/0.log" Oct 03 14:19:14 crc kubenswrapper[4578]: I1003 14:19:14.427194 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-479g8_5d79d7a8-f4ba-4f27-8860-97c9d8f36b8f/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 14:19:14 crc kubenswrapper[4578]: I1003 14:19:14.599525 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-ht44x_a12dea99-870e-4d7b-81e2-f659f1a3c0ee/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 14:19:14 crc kubenswrapper[4578]: I1003 14:19:14.908469 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-457zq_6d166384-568f-42ea-a441-4d1df54fd5ce/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 14:19:15 crc kubenswrapper[4578]: I1003 14:19:15.171367 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-nmz46_187edc5e-0ec9-4d8f-8085-2b0a393df59a/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 14:19:15 crc kubenswrapper[4578]: I1003 14:19:15.222221 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-ggkhk_ac850868-8034-40e9-ac47-21455b1f8817/ssh-known-hosts-edpm-deployment/0.log" Oct 03 14:19:15 crc kubenswrapper[4578]: I1003 14:19:15.708089 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-57dc876bc5-gvzk7_bd42e3ab-aa85-453e-abfd-7b6aa55e9674/proxy-server/0.log" Oct 03 14:19:15 crc kubenswrapper[4578]: I1003 14:19:15.778496 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-57dc876bc5-gvzk7_bd42e3ab-aa85-453e-abfd-7b6aa55e9674/proxy-httpd/0.log" Oct 03 14:19:15 crc kubenswrapper[4578]: I1003 14:19:15.977822 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-hjk2v_3f2222d5-bfbb-43ca-a40a-d235295823a8/swift-ring-rebalance/0.log" Oct 03 14:19:16 crc kubenswrapper[4578]: I1003 14:19:16.210399 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_aa1d7b22-8dac-42de-8241-b633bbe90691/account-auditor/0.log" Oct 03 14:19:16 crc kubenswrapper[4578]: I1003 14:19:16.247426 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_aa1d7b22-8dac-42de-8241-b633bbe90691/account-reaper/0.log" Oct 03 14:19:16 crc kubenswrapper[4578]: I1003 14:19:16.969146 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_aa1d7b22-8dac-42de-8241-b633bbe90691/account-server/0.log" Oct 03 14:19:17 crc kubenswrapper[4578]: I1003 14:19:17.052369 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_aa1d7b22-8dac-42de-8241-b633bbe90691/container-auditor/0.log" Oct 03 14:19:17 crc kubenswrapper[4578]: I1003 14:19:17.059735 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_aa1d7b22-8dac-42de-8241-b633bbe90691/account-replicator/0.log" Oct 03 14:19:17 crc kubenswrapper[4578]: I1003 14:19:17.256017 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_aa1d7b22-8dac-42de-8241-b633bbe90691/container-replicator/0.log" Oct 03 14:19:17 crc kubenswrapper[4578]: I1003 14:19:17.323854 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_aa1d7b22-8dac-42de-8241-b633bbe90691/container-server/0.log" Oct 03 14:19:17 crc kubenswrapper[4578]: I1003 14:19:17.336040 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_aa1d7b22-8dac-42de-8241-b633bbe90691/container-updater/0.log" Oct 03 14:19:17 crc kubenswrapper[4578]: I1003 14:19:17.605054 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_aa1d7b22-8dac-42de-8241-b633bbe90691/object-expirer/0.log" Oct 03 14:19:17 crc kubenswrapper[4578]: I1003 14:19:17.702148 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_aa1d7b22-8dac-42de-8241-b633bbe90691/object-replicator/0.log" Oct 03 14:19:17 crc kubenswrapper[4578]: I1003 14:19:17.703565 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_aa1d7b22-8dac-42de-8241-b633bbe90691/object-auditor/0.log" Oct 03 14:19:17 crc kubenswrapper[4578]: I1003 14:19:17.750560 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_409ed574-cc6d-49e4-b1b1-4ec50dfe171e/memcached/0.log" Oct 03 14:19:17 crc kubenswrapper[4578]: I1003 14:19:17.922284 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_aa1d7b22-8dac-42de-8241-b633bbe90691/object-server/0.log" Oct 03 14:19:17 crc kubenswrapper[4578]: I1003 14:19:17.953760 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_aa1d7b22-8dac-42de-8241-b633bbe90691/rsync/0.log" Oct 03 14:19:17 crc kubenswrapper[4578]: I1003 14:19:17.984359 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_aa1d7b22-8dac-42de-8241-b633bbe90691/object-updater/0.log" Oct 03 14:19:18 crc kubenswrapper[4578]: I1003 14:19:18.018484 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_aa1d7b22-8dac-42de-8241-b633bbe90691/swift-recon-cron/0.log" Oct 03 14:19:18 crc kubenswrapper[4578]: I1003 14:19:18.165887 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-dcxd9_7a2213e2-d0c1-4481-898e-c1d9a52164d6/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 03 14:19:39 crc kubenswrapper[4578]: I1003 14:19:39.575852 4578 generic.go:334] "Generic (PLEG): container finished" podID="0d5b463e-66dd-4d15-8713-6a762ef6fdc4" containerID="4269a4450af0f08820987b8080dd0e13e068a1c7afa451e8a8bd73bb1476c239" exitCode=0 Oct 03 14:19:39 crc kubenswrapper[4578]: I1003 14:19:39.575920 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mcgpg/crc-debug-jjz6m" event={"ID":"0d5b463e-66dd-4d15-8713-6a762ef6fdc4","Type":"ContainerDied","Data":"4269a4450af0f08820987b8080dd0e13e068a1c7afa451e8a8bd73bb1476c239"} Oct 03 14:19:40 crc kubenswrapper[4578]: I1003 14:19:40.702394 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mcgpg/crc-debug-jjz6m" Oct 03 14:19:40 crc kubenswrapper[4578]: I1003 14:19:40.736057 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mcgpg/crc-debug-jjz6m"] Oct 03 14:19:40 crc kubenswrapper[4578]: I1003 14:19:40.743258 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mcgpg/crc-debug-jjz6m"] Oct 03 14:19:40 crc kubenswrapper[4578]: I1003 14:19:40.778593 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d5b463e-66dd-4d15-8713-6a762ef6fdc4-host\") pod \"0d5b463e-66dd-4d15-8713-6a762ef6fdc4\" (UID: \"0d5b463e-66dd-4d15-8713-6a762ef6fdc4\") " Oct 03 14:19:40 crc kubenswrapper[4578]: I1003 14:19:40.778764 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d5b463e-66dd-4d15-8713-6a762ef6fdc4-host" (OuterVolumeSpecName: "host") pod "0d5b463e-66dd-4d15-8713-6a762ef6fdc4" (UID: "0d5b463e-66dd-4d15-8713-6a762ef6fdc4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 14:19:40 crc kubenswrapper[4578]: I1003 14:19:40.778892 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxqxh\" (UniqueName: \"kubernetes.io/projected/0d5b463e-66dd-4d15-8713-6a762ef6fdc4-kube-api-access-qxqxh\") pod \"0d5b463e-66dd-4d15-8713-6a762ef6fdc4\" (UID: \"0d5b463e-66dd-4d15-8713-6a762ef6fdc4\") " Oct 03 14:19:40 crc kubenswrapper[4578]: I1003 14:19:40.779382 4578 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d5b463e-66dd-4d15-8713-6a762ef6fdc4-host\") on node \"crc\" DevicePath \"\"" Oct 03 14:19:40 crc kubenswrapper[4578]: I1003 14:19:40.785868 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d5b463e-66dd-4d15-8713-6a762ef6fdc4-kube-api-access-qxqxh" (OuterVolumeSpecName: "kube-api-access-qxqxh") pod "0d5b463e-66dd-4d15-8713-6a762ef6fdc4" (UID: "0d5b463e-66dd-4d15-8713-6a762ef6fdc4"). InnerVolumeSpecName "kube-api-access-qxqxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:19:40 crc kubenswrapper[4578]: I1003 14:19:40.881374 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxqxh\" (UniqueName: \"kubernetes.io/projected/0d5b463e-66dd-4d15-8713-6a762ef6fdc4-kube-api-access-qxqxh\") on node \"crc\" DevicePath \"\"" Oct 03 14:19:40 crc kubenswrapper[4578]: I1003 14:19:40.920257 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d5b463e-66dd-4d15-8713-6a762ef6fdc4" path="/var/lib/kubelet/pods/0d5b463e-66dd-4d15-8713-6a762ef6fdc4/volumes" Oct 03 14:19:41 crc kubenswrapper[4578]: I1003 14:19:41.594044 4578 scope.go:117] "RemoveContainer" containerID="4269a4450af0f08820987b8080dd0e13e068a1c7afa451e8a8bd73bb1476c239" Oct 03 14:19:41 crc kubenswrapper[4578]: I1003 14:19:41.594081 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mcgpg/crc-debug-jjz6m" Oct 03 14:19:41 crc kubenswrapper[4578]: I1003 14:19:41.935543 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mcgpg/crc-debug-fpxkr"] Oct 03 14:19:41 crc kubenswrapper[4578]: E1003 14:19:41.936280 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d5b463e-66dd-4d15-8713-6a762ef6fdc4" containerName="container-00" Oct 03 14:19:41 crc kubenswrapper[4578]: I1003 14:19:41.936312 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d5b463e-66dd-4d15-8713-6a762ef6fdc4" containerName="container-00" Oct 03 14:19:41 crc kubenswrapper[4578]: I1003 14:19:41.936568 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d5b463e-66dd-4d15-8713-6a762ef6fdc4" containerName="container-00" Oct 03 14:19:41 crc kubenswrapper[4578]: I1003 14:19:41.937412 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mcgpg/crc-debug-fpxkr" Oct 03 14:19:41 crc kubenswrapper[4578]: I1003 14:19:41.999647 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/34f2cceb-0248-4b41-91d4-3eddf089ee7b-host\") pod \"crc-debug-fpxkr\" (UID: \"34f2cceb-0248-4b41-91d4-3eddf089ee7b\") " pod="openshift-must-gather-mcgpg/crc-debug-fpxkr" Oct 03 14:19:41 crc kubenswrapper[4578]: I1003 14:19:41.999748 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl9nn\" (UniqueName: \"kubernetes.io/projected/34f2cceb-0248-4b41-91d4-3eddf089ee7b-kube-api-access-tl9nn\") pod \"crc-debug-fpxkr\" (UID: \"34f2cceb-0248-4b41-91d4-3eddf089ee7b\") " pod="openshift-must-gather-mcgpg/crc-debug-fpxkr" Oct 03 14:19:42 crc kubenswrapper[4578]: I1003 14:19:42.101712 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/34f2cceb-0248-4b41-91d4-3eddf089ee7b-host\") pod \"crc-debug-fpxkr\" (UID: \"34f2cceb-0248-4b41-91d4-3eddf089ee7b\") " pod="openshift-must-gather-mcgpg/crc-debug-fpxkr" Oct 03 14:19:42 crc kubenswrapper[4578]: I1003 14:19:42.101806 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl9nn\" (UniqueName: \"kubernetes.io/projected/34f2cceb-0248-4b41-91d4-3eddf089ee7b-kube-api-access-tl9nn\") pod \"crc-debug-fpxkr\" (UID: \"34f2cceb-0248-4b41-91d4-3eddf089ee7b\") " pod="openshift-must-gather-mcgpg/crc-debug-fpxkr" Oct 03 14:19:42 crc kubenswrapper[4578]: I1003 14:19:42.101899 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/34f2cceb-0248-4b41-91d4-3eddf089ee7b-host\") pod \"crc-debug-fpxkr\" (UID: \"34f2cceb-0248-4b41-91d4-3eddf089ee7b\") " pod="openshift-must-gather-mcgpg/crc-debug-fpxkr" Oct 03 14:19:42 crc kubenswrapper[4578]: I1003 14:19:42.121553 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl9nn\" (UniqueName: \"kubernetes.io/projected/34f2cceb-0248-4b41-91d4-3eddf089ee7b-kube-api-access-tl9nn\") pod \"crc-debug-fpxkr\" (UID: \"34f2cceb-0248-4b41-91d4-3eddf089ee7b\") " pod="openshift-must-gather-mcgpg/crc-debug-fpxkr" Oct 03 14:19:42 crc kubenswrapper[4578]: I1003 14:19:42.270278 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mcgpg/crc-debug-fpxkr" Oct 03 14:19:42 crc kubenswrapper[4578]: I1003 14:19:42.604669 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mcgpg/crc-debug-fpxkr" event={"ID":"34f2cceb-0248-4b41-91d4-3eddf089ee7b","Type":"ContainerStarted","Data":"a4429f2e6d178f4cbfa5b1168ef9d2d7cb8459999b810cc396981aa3d84936d1"} Oct 03 14:19:43 crc kubenswrapper[4578]: I1003 14:19:43.616124 4578 generic.go:334] "Generic (PLEG): container finished" podID="34f2cceb-0248-4b41-91d4-3eddf089ee7b" containerID="7bee42350cf8e03b9481385b9499627f52e748a1353e34719e10a84720d93f67" exitCode=0 Oct 03 14:19:43 crc kubenswrapper[4578]: I1003 14:19:43.616224 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mcgpg/crc-debug-fpxkr" event={"ID":"34f2cceb-0248-4b41-91d4-3eddf089ee7b","Type":"ContainerDied","Data":"7bee42350cf8e03b9481385b9499627f52e748a1353e34719e10a84720d93f67"} Oct 03 14:19:44 crc kubenswrapper[4578]: I1003 14:19:44.731861 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mcgpg/crc-debug-fpxkr" Oct 03 14:19:44 crc kubenswrapper[4578]: I1003 14:19:44.848360 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tl9nn\" (UniqueName: \"kubernetes.io/projected/34f2cceb-0248-4b41-91d4-3eddf089ee7b-kube-api-access-tl9nn\") pod \"34f2cceb-0248-4b41-91d4-3eddf089ee7b\" (UID: \"34f2cceb-0248-4b41-91d4-3eddf089ee7b\") " Oct 03 14:19:44 crc kubenswrapper[4578]: I1003 14:19:44.848510 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/34f2cceb-0248-4b41-91d4-3eddf089ee7b-host\") pod \"34f2cceb-0248-4b41-91d4-3eddf089ee7b\" (UID: \"34f2cceb-0248-4b41-91d4-3eddf089ee7b\") " Oct 03 14:19:44 crc kubenswrapper[4578]: I1003 14:19:44.848838 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34f2cceb-0248-4b41-91d4-3eddf089ee7b-host" (OuterVolumeSpecName: "host") pod "34f2cceb-0248-4b41-91d4-3eddf089ee7b" (UID: "34f2cceb-0248-4b41-91d4-3eddf089ee7b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 14:19:44 crc kubenswrapper[4578]: I1003 14:19:44.849170 4578 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/34f2cceb-0248-4b41-91d4-3eddf089ee7b-host\") on node \"crc\" DevicePath \"\"" Oct 03 14:19:44 crc kubenswrapper[4578]: I1003 14:19:44.875405 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34f2cceb-0248-4b41-91d4-3eddf089ee7b-kube-api-access-tl9nn" (OuterVolumeSpecName: "kube-api-access-tl9nn") pod "34f2cceb-0248-4b41-91d4-3eddf089ee7b" (UID: "34f2cceb-0248-4b41-91d4-3eddf089ee7b"). InnerVolumeSpecName "kube-api-access-tl9nn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:19:44 crc kubenswrapper[4578]: I1003 14:19:44.950391 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tl9nn\" (UniqueName: \"kubernetes.io/projected/34f2cceb-0248-4b41-91d4-3eddf089ee7b-kube-api-access-tl9nn\") on node \"crc\" DevicePath \"\"" Oct 03 14:19:45 crc kubenswrapper[4578]: I1003 14:19:45.643275 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mcgpg/crc-debug-fpxkr" event={"ID":"34f2cceb-0248-4b41-91d4-3eddf089ee7b","Type":"ContainerDied","Data":"a4429f2e6d178f4cbfa5b1168ef9d2d7cb8459999b810cc396981aa3d84936d1"} Oct 03 14:19:45 crc kubenswrapper[4578]: I1003 14:19:45.643602 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4429f2e6d178f4cbfa5b1168ef9d2d7cb8459999b810cc396981aa3d84936d1" Oct 03 14:19:45 crc kubenswrapper[4578]: I1003 14:19:45.643455 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mcgpg/crc-debug-fpxkr" Oct 03 14:19:49 crc kubenswrapper[4578]: I1003 14:19:49.886048 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mcgpg/crc-debug-fpxkr"] Oct 03 14:19:49 crc kubenswrapper[4578]: I1003 14:19:49.899669 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mcgpg/crc-debug-fpxkr"] Oct 03 14:19:50 crc kubenswrapper[4578]: I1003 14:19:50.920135 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34f2cceb-0248-4b41-91d4-3eddf089ee7b" path="/var/lib/kubelet/pods/34f2cceb-0248-4b41-91d4-3eddf089ee7b/volumes" Oct 03 14:19:51 crc kubenswrapper[4578]: I1003 14:19:51.087239 4578 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mcgpg/crc-debug-s8z72"] Oct 03 14:19:51 crc kubenswrapper[4578]: E1003 14:19:51.087883 4578 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34f2cceb-0248-4b41-91d4-3eddf089ee7b" containerName="container-00" Oct 03 14:19:51 crc kubenswrapper[4578]: I1003 14:19:51.087902 4578 state_mem.go:107] "Deleted CPUSet assignment" podUID="34f2cceb-0248-4b41-91d4-3eddf089ee7b" containerName="container-00" Oct 03 14:19:51 crc kubenswrapper[4578]: I1003 14:19:51.088114 4578 memory_manager.go:354] "RemoveStaleState removing state" podUID="34f2cceb-0248-4b41-91d4-3eddf089ee7b" containerName="container-00" Oct 03 14:19:51 crc kubenswrapper[4578]: I1003 14:19:51.088737 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mcgpg/crc-debug-s8z72" Oct 03 14:19:51 crc kubenswrapper[4578]: I1003 14:19:51.190852 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn96d\" (UniqueName: \"kubernetes.io/projected/905d4cfd-be2d-423b-b741-66aaf63def45-kube-api-access-xn96d\") pod \"crc-debug-s8z72\" (UID: \"905d4cfd-be2d-423b-b741-66aaf63def45\") " pod="openshift-must-gather-mcgpg/crc-debug-s8z72" Oct 03 14:19:51 crc kubenswrapper[4578]: I1003 14:19:51.190972 4578 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/905d4cfd-be2d-423b-b741-66aaf63def45-host\") pod \"crc-debug-s8z72\" (UID: \"905d4cfd-be2d-423b-b741-66aaf63def45\") " pod="openshift-must-gather-mcgpg/crc-debug-s8z72" Oct 03 14:19:51 crc kubenswrapper[4578]: I1003 14:19:51.292377 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn96d\" (UniqueName: \"kubernetes.io/projected/905d4cfd-be2d-423b-b741-66aaf63def45-kube-api-access-xn96d\") pod \"crc-debug-s8z72\" (UID: \"905d4cfd-be2d-423b-b741-66aaf63def45\") " pod="openshift-must-gather-mcgpg/crc-debug-s8z72" Oct 03 14:19:51 crc kubenswrapper[4578]: I1003 14:19:51.292461 4578 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/905d4cfd-be2d-423b-b741-66aaf63def45-host\") pod \"crc-debug-s8z72\" (UID: \"905d4cfd-be2d-423b-b741-66aaf63def45\") " pod="openshift-must-gather-mcgpg/crc-debug-s8z72" Oct 03 14:19:51 crc kubenswrapper[4578]: I1003 14:19:51.292651 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/905d4cfd-be2d-423b-b741-66aaf63def45-host\") pod \"crc-debug-s8z72\" (UID: \"905d4cfd-be2d-423b-b741-66aaf63def45\") " pod="openshift-must-gather-mcgpg/crc-debug-s8z72" Oct 03 14:19:51 crc kubenswrapper[4578]: I1003 14:19:51.316811 4578 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn96d\" (UniqueName: \"kubernetes.io/projected/905d4cfd-be2d-423b-b741-66aaf63def45-kube-api-access-xn96d\") pod \"crc-debug-s8z72\" (UID: \"905d4cfd-be2d-423b-b741-66aaf63def45\") " pod="openshift-must-gather-mcgpg/crc-debug-s8z72" Oct 03 14:19:51 crc kubenswrapper[4578]: I1003 14:19:51.407156 4578 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mcgpg/crc-debug-s8z72" Oct 03 14:19:51 crc kubenswrapper[4578]: I1003 14:19:51.696220 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mcgpg/crc-debug-s8z72" event={"ID":"905d4cfd-be2d-423b-b741-66aaf63def45","Type":"ContainerStarted","Data":"34e8fdfa13d17bac7782ed3cc0eafb932a396a6dab5ce57e6edd16412476c2dd"} Oct 03 14:19:51 crc kubenswrapper[4578]: I1003 14:19:51.696434 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mcgpg/crc-debug-s8z72" event={"ID":"905d4cfd-be2d-423b-b741-66aaf63def45","Type":"ContainerStarted","Data":"7fe973bdf7c5f16a258332fe6809598e7e031e1f88da0df6dda3b61de6710fd2"} Oct 03 14:19:51 crc kubenswrapper[4578]: I1003 14:19:51.721779 4578 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mcgpg/crc-debug-s8z72" podStartSLOduration=0.721761917 podStartE2EDuration="721.761917ms" podCreationTimestamp="2025-10-03 14:19:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-03 14:19:51.711698704 +0000 UTC m=+5327.510170908" watchObservedRunningTime="2025-10-03 14:19:51.721761917 +0000 UTC m=+5327.520234101" Oct 03 14:19:52 crc kubenswrapper[4578]: I1003 14:19:52.711597 4578 generic.go:334] "Generic (PLEG): container finished" podID="905d4cfd-be2d-423b-b741-66aaf63def45" containerID="34e8fdfa13d17bac7782ed3cc0eafb932a396a6dab5ce57e6edd16412476c2dd" exitCode=0 Oct 03 14:19:52 crc kubenswrapper[4578]: I1003 14:19:52.711863 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mcgpg/crc-debug-s8z72" event={"ID":"905d4cfd-be2d-423b-b741-66aaf63def45","Type":"ContainerDied","Data":"34e8fdfa13d17bac7782ed3cc0eafb932a396a6dab5ce57e6edd16412476c2dd"} Oct 03 14:19:53 crc kubenswrapper[4578]: I1003 14:19:53.815180 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mcgpg/crc-debug-s8z72" Oct 03 14:19:53 crc kubenswrapper[4578]: I1003 14:19:53.846832 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mcgpg/crc-debug-s8z72"] Oct 03 14:19:53 crc kubenswrapper[4578]: I1003 14:19:53.853162 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mcgpg/crc-debug-s8z72"] Oct 03 14:19:53 crc kubenswrapper[4578]: I1003 14:19:53.948601 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/905d4cfd-be2d-423b-b741-66aaf63def45-host\") pod \"905d4cfd-be2d-423b-b741-66aaf63def45\" (UID: \"905d4cfd-be2d-423b-b741-66aaf63def45\") " Oct 03 14:19:53 crc kubenswrapper[4578]: I1003 14:19:53.949096 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xn96d\" (UniqueName: \"kubernetes.io/projected/905d4cfd-be2d-423b-b741-66aaf63def45-kube-api-access-xn96d\") pod \"905d4cfd-be2d-423b-b741-66aaf63def45\" (UID: \"905d4cfd-be2d-423b-b741-66aaf63def45\") " Oct 03 14:19:53 crc kubenswrapper[4578]: I1003 14:19:53.953506 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/905d4cfd-be2d-423b-b741-66aaf63def45-host" (OuterVolumeSpecName: "host") pod "905d4cfd-be2d-423b-b741-66aaf63def45" (UID: "905d4cfd-be2d-423b-b741-66aaf63def45"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 03 14:19:53 crc kubenswrapper[4578]: I1003 14:19:53.960280 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/905d4cfd-be2d-423b-b741-66aaf63def45-kube-api-access-xn96d" (OuterVolumeSpecName: "kube-api-access-xn96d") pod "905d4cfd-be2d-423b-b741-66aaf63def45" (UID: "905d4cfd-be2d-423b-b741-66aaf63def45"). InnerVolumeSpecName "kube-api-access-xn96d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:19:54 crc kubenswrapper[4578]: I1003 14:19:54.052165 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xn96d\" (UniqueName: \"kubernetes.io/projected/905d4cfd-be2d-423b-b741-66aaf63def45-kube-api-access-xn96d\") on node \"crc\" DevicePath \"\"" Oct 03 14:19:54 crc kubenswrapper[4578]: I1003 14:19:54.052205 4578 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/905d4cfd-be2d-423b-b741-66aaf63def45-host\") on node \"crc\" DevicePath \"\"" Oct 03 14:19:54 crc kubenswrapper[4578]: I1003 14:19:54.731260 4578 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fe973bdf7c5f16a258332fe6809598e7e031e1f88da0df6dda3b61de6710fd2" Oct 03 14:19:54 crc kubenswrapper[4578]: I1003 14:19:54.731502 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mcgpg/crc-debug-s8z72" Oct 03 14:19:54 crc kubenswrapper[4578]: I1003 14:19:54.919250 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="905d4cfd-be2d-423b-b741-66aaf63def45" path="/var/lib/kubelet/pods/905d4cfd-be2d-423b-b741-66aaf63def45/volumes" Oct 03 14:19:55 crc kubenswrapper[4578]: I1003 14:19:55.520713 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9_cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d/util/0.log" Oct 03 14:19:55 crc kubenswrapper[4578]: I1003 14:19:55.761011 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9_cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d/util/0.log" Oct 03 14:19:55 crc kubenswrapper[4578]: I1003 14:19:55.828770 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9_cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d/pull/0.log" Oct 03 14:19:55 crc kubenswrapper[4578]: I1003 14:19:55.863786 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9_cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d/pull/0.log" Oct 03 14:19:55 crc kubenswrapper[4578]: I1003 14:19:55.969399 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9_cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d/pull/0.log" Oct 03 14:19:56 crc kubenswrapper[4578]: I1003 14:19:56.085180 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9_cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d/util/0.log" Oct 03 14:19:56 crc kubenswrapper[4578]: I1003 14:19:56.391244 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6866606fa3a289e0b44cd13ac7038d6356f0a6aa62e0445808c76e969asg7k9_cd645ad4-0d57-4dc2-9b0b-a6e025a5be5d/extract/0.log" Oct 03 14:19:56 crc kubenswrapper[4578]: I1003 14:19:56.440049 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6c675fb79f-8sdmv_b9f0875d-37d8-4486-8618-eb0ff333f1d5/kube-rbac-proxy/0.log" Oct 03 14:19:56 crc kubenswrapper[4578]: I1003 14:19:56.582467 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79d68d6c85-jb6mh_2d6b7a4e-dd13-443e-ac6b-bc0882a0a773/kube-rbac-proxy/0.log" Oct 03 14:19:56 crc kubenswrapper[4578]: I1003 14:19:56.709028 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6c675fb79f-8sdmv_b9f0875d-37d8-4486-8618-eb0ff333f1d5/manager/0.log" Oct 03 14:19:56 crc kubenswrapper[4578]: I1003 14:19:56.731850 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79d68d6c85-jb6mh_2d6b7a4e-dd13-443e-ac6b-bc0882a0a773/manager/0.log" Oct 03 14:19:56 crc kubenswrapper[4578]: I1003 14:19:56.886064 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-x8jtc_b1fc462d-d689-4a5a-b0cb-d71ea7cb8f76/kube-rbac-proxy/0.log" Oct 03 14:19:56 crc kubenswrapper[4578]: I1003 14:19:56.931828 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-x8jtc_b1fc462d-d689-4a5a-b0cb-d71ea7cb8f76/manager/0.log" Oct 03 14:19:57 crc kubenswrapper[4578]: I1003 14:19:57.077992 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-846dff85b5-q5cs2_86ba4ceb-a98b-473f-a133-c20b0d95ca04/kube-rbac-proxy/0.log" Oct 03 14:19:57 crc kubenswrapper[4578]: I1003 14:19:57.231881 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-846dff85b5-q5cs2_86ba4ceb-a98b-473f-a133-c20b0d95ca04/manager/0.log" Oct 03 14:19:57 crc kubenswrapper[4578]: I1003 14:19:57.284537 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-599898f689-vjh6c_311cfa36-937f-4531-ab2e-f4dec6164051/manager/0.log" Oct 03 14:19:57 crc kubenswrapper[4578]: I1003 14:19:57.294506 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-599898f689-vjh6c_311cfa36-937f-4531-ab2e-f4dec6164051/kube-rbac-proxy/0.log" Oct 03 14:19:57 crc kubenswrapper[4578]: I1003 14:19:57.473664 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6769b867d9-dt8s5_86baff4f-cbb5-470f-919c-d8b571a8befb/kube-rbac-proxy/0.log" Oct 03 14:19:57 crc kubenswrapper[4578]: I1003 14:19:57.558423 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6769b867d9-dt8s5_86baff4f-cbb5-470f-919c-d8b571a8befb/manager/0.log" Oct 03 14:19:57 crc kubenswrapper[4578]: I1003 14:19:57.708545 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5fbf469cd7-d6mqh_a517e8b1-3fa5-41f6-a6a5-d22d9f341201/kube-rbac-proxy/0.log" Oct 03 14:19:57 crc kubenswrapper[4578]: I1003 14:19:57.840744 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5fbf469cd7-d6mqh_a517e8b1-3fa5-41f6-a6a5-d22d9f341201/manager/0.log" Oct 03 14:19:57 crc kubenswrapper[4578]: I1003 14:19:57.891540 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-84bc9db6cc-2f2dh_f6d68890-a075-4dca-b4cf-19325838c18e/kube-rbac-proxy/0.log" Oct 03 14:19:58 crc kubenswrapper[4578]: I1003 14:19:58.054486 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-84bc9db6cc-2f2dh_f6d68890-a075-4dca-b4cf-19325838c18e/manager/0.log" Oct 03 14:19:58 crc kubenswrapper[4578]: I1003 14:19:58.079009 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7f55849f88-92tt9_1e453068-d56e-4085-85fa-3bbaadd48c13/kube-rbac-proxy/0.log" Oct 03 14:19:58 crc kubenswrapper[4578]: I1003 14:19:58.258340 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7f55849f88-92tt9_1e453068-d56e-4085-85fa-3bbaadd48c13/manager/0.log" Oct 03 14:19:58 crc kubenswrapper[4578]: I1003 14:19:58.303660 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6fd6854b49-kgt86_efbb9c56-e481-4f3e-b657-a97bd2952eee/kube-rbac-proxy/0.log" Oct 03 14:19:58 crc kubenswrapper[4578]: I1003 14:19:58.318345 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6fd6854b49-kgt86_efbb9c56-e481-4f3e-b657-a97bd2952eee/manager/0.log" Oct 03 14:19:58 crc kubenswrapper[4578]: I1003 14:19:58.468189 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5c468bf4d4-cn6bd_e252fd2f-b318-4292-b359-7b42ca159c26/kube-rbac-proxy/0.log" Oct 03 14:19:58 crc kubenswrapper[4578]: I1003 14:19:58.602020 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5c468bf4d4-cn6bd_e252fd2f-b318-4292-b359-7b42ca159c26/manager/0.log" Oct 03 14:19:58 crc kubenswrapper[4578]: I1003 14:19:58.711683 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6574bf987d-s54bp_249acba6-7e64-4f9f-8f60-22740371de34/kube-rbac-proxy/0.log" Oct 03 14:19:58 crc kubenswrapper[4578]: I1003 14:19:58.854293 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6574bf987d-s54bp_249acba6-7e64-4f9f-8f60-22740371de34/manager/0.log" Oct 03 14:19:58 crc kubenswrapper[4578]: I1003 14:19:58.946853 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-555c7456bd-p4chq_5848e7b5-08c6-4a7d-9fe9-3021f8aa0ac8/kube-rbac-proxy/0.log" Oct 03 14:19:59 crc kubenswrapper[4578]: I1003 14:19:59.095349 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-555c7456bd-p4chq_5848e7b5-08c6-4a7d-9fe9-3021f8aa0ac8/manager/0.log" Oct 03 14:19:59 crc kubenswrapper[4578]: I1003 14:19:59.166893 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-59d6cfdf45-b28kv_a152b6b9-86b3-489d-9ceb-a9a284e054d1/kube-rbac-proxy/0.log" Oct 03 14:19:59 crc kubenswrapper[4578]: I1003 14:19:59.195433 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-59d6cfdf45-b28kv_a152b6b9-86b3-489d-9ceb-a9a284e054d1/manager/0.log" Oct 03 14:19:59 crc kubenswrapper[4578]: I1003 14:19:59.385939 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6f64c4d6789s9md_31250d8b-75f5-4145-b8ed-1ec79bebe936/kube-rbac-proxy/0.log" Oct 03 14:19:59 crc kubenswrapper[4578]: I1003 14:19:59.452960 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6f64c4d6789s9md_31250d8b-75f5-4145-b8ed-1ec79bebe936/manager/0.log" Oct 03 14:19:59 crc kubenswrapper[4578]: I1003 14:19:59.604388 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5c4446bf96-nxg6g_4855be16-97a1-467a-bfbe-36b2f15a5afd/kube-rbac-proxy/0.log" Oct 03 14:19:59 crc kubenswrapper[4578]: I1003 14:19:59.697807 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-764f84468b-x95lq_440efe1c-5d5c-4e97-a684-3afa7a78b8d7/kube-rbac-proxy/0.log" Oct 03 14:19:59 crc kubenswrapper[4578]: I1003 14:19:59.965277 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-svcqt_49c934f0-4e15-40cf-a0e6-f067f39912f7/registry-server/0.log" Oct 03 14:19:59 crc kubenswrapper[4578]: I1003 14:19:59.968496 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-764f84468b-x95lq_440efe1c-5d5c-4e97-a684-3afa7a78b8d7/operator/0.log" Oct 03 14:20:00 crc kubenswrapper[4578]: I1003 14:20:00.197104 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-688db7b6c7-6ft9g_48ea8320-2f7a-46e5-8681-73d075a6b74b/kube-rbac-proxy/0.log" Oct 03 14:20:00 crc kubenswrapper[4578]: I1003 14:20:00.345815 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-688db7b6c7-6ft9g_48ea8320-2f7a-46e5-8681-73d075a6b74b/manager/0.log" Oct 03 14:20:00 crc kubenswrapper[4578]: I1003 14:20:00.436933 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-7d8bb7f44c-ql5zc_d8bb1592-732e-4e22-944e-3298b6937f55/kube-rbac-proxy/0.log" Oct 03 14:20:00 crc kubenswrapper[4578]: I1003 14:20:00.591734 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-7d8bb7f44c-ql5zc_d8bb1592-732e-4e22-944e-3298b6937f55/manager/0.log" Oct 03 14:20:00 crc kubenswrapper[4578]: I1003 14:20:00.688540 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-mj2n5_1141b204-4370-4c24-bbf3-71734e237805/operator/0.log" Oct 03 14:20:00 crc kubenswrapper[4578]: I1003 14:20:00.773520 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5c4446bf96-nxg6g_4855be16-97a1-467a-bfbe-36b2f15a5afd/manager/0.log" Oct 03 14:20:00 crc kubenswrapper[4578]: I1003 14:20:00.897365 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-kzr6z_65e1c76d-7bb5-4c09-8b8b-d54f3d92c42b/kube-rbac-proxy/0.log" Oct 03 14:20:00 crc kubenswrapper[4578]: I1003 14:20:00.932816 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-kzr6z_65e1c76d-7bb5-4c09-8b8b-d54f3d92c42b/manager/0.log" Oct 03 14:20:01 crc kubenswrapper[4578]: I1003 14:20:01.051654 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5db5cf686f-rpzgw_1a74246b-d89b-424f-bd47-d57674b06855/kube-rbac-proxy/0.log" Oct 03 14:20:01 crc kubenswrapper[4578]: I1003 14:20:01.059115 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5db5cf686f-rpzgw_1a74246b-d89b-424f-bd47-d57674b06855/manager/0.log" Oct 03 14:20:01 crc kubenswrapper[4578]: I1003 14:20:01.139188 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-r2zgb_5b7d3469-bfd4-47c8-a82c-3e00721b2102/kube-rbac-proxy/0.log" Oct 03 14:20:01 crc kubenswrapper[4578]: I1003 14:20:01.221439 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-r2zgb_5b7d3469-bfd4-47c8-a82c-3e00721b2102/manager/0.log" Oct 03 14:20:01 crc kubenswrapper[4578]: I1003 14:20:01.255363 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-fcd7d9895-dcr4b_4a1657ed-b0b2-4f52-9f9c-38ed6d6bb8ab/kube-rbac-proxy/0.log" Oct 03 14:20:01 crc kubenswrapper[4578]: I1003 14:20:01.371444 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-fcd7d9895-dcr4b_4a1657ed-b0b2-4f52-9f9c-38ed6d6bb8ab/manager/0.log" Oct 03 14:20:16 crc kubenswrapper[4578]: I1003 14:20:16.548299 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-hpc5d_6f3f593c-f784-4f1c-8fb3-016275c38649/control-plane-machine-set-operator/0.log" Oct 03 14:20:16 crc kubenswrapper[4578]: I1003 14:20:16.690062 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zsb7g_2f691127-6990-46bd-afa7-d6f3f7eaa329/kube-rbac-proxy/0.log" Oct 03 14:20:16 crc kubenswrapper[4578]: I1003 14:20:16.741889 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-zsb7g_2f691127-6990-46bd-afa7-d6f3f7eaa329/machine-api-operator/0.log" Oct 03 14:20:27 crc kubenswrapper[4578]: I1003 14:20:27.618315 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-rb7jq_7350f95d-fefa-47f4-85fb-9ac7a9753bec/cert-manager-controller/0.log" Oct 03 14:20:27 crc kubenswrapper[4578]: I1003 14:20:27.705254 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-d4qcd_fde8ee73-946e-4c94-85ff-34f54a387575/cert-manager-cainjector/0.log" Oct 03 14:20:27 crc kubenswrapper[4578]: I1003 14:20:27.827648 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-xp9mm_5cb84a0f-4490-4370-b07e-71ecb650fc07/cert-manager-webhook/0.log" Oct 03 14:20:39 crc kubenswrapper[4578]: I1003 14:20:39.288267 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-4xjn2_a33377f4-f20e-42e4-af02-1f9a7bffc587/nmstate-console-plugin/0.log" Oct 03 14:20:39 crc kubenswrapper[4578]: I1003 14:20:39.476732 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-652tz_94dd97ab-9637-4eab-82dc-5d070129abac/kube-rbac-proxy/0.log" Oct 03 14:20:39 crc kubenswrapper[4578]: I1003 14:20:39.502171 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-vknp5_b111b867-4d61-489f-aa89-f820b50c9a92/nmstate-handler/0.log" Oct 03 14:20:39 crc kubenswrapper[4578]: I1003 14:20:39.611792 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-652tz_94dd97ab-9637-4eab-82dc-5d070129abac/nmstate-metrics/0.log" Oct 03 14:20:39 crc kubenswrapper[4578]: I1003 14:20:39.697989 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-54fwn_566708f1-bb6e-4f0b-b381-d594f60b864b/nmstate-operator/0.log" Oct 03 14:20:39 crc kubenswrapper[4578]: I1003 14:20:39.878439 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-fp4rx_bf9f48d4-44a0-4a43-872d-fce41608a68e/nmstate-webhook/0.log" Oct 03 14:20:55 crc kubenswrapper[4578]: I1003 14:20:55.489005 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-jhpgk_28ab5270-7890-40e5-8414-fd4a1b9d2742/kube-rbac-proxy/0.log" Oct 03 14:20:55 crc kubenswrapper[4578]: I1003 14:20:55.565462 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-jhpgk_28ab5270-7890-40e5-8414-fd4a1b9d2742/controller/0.log" Oct 03 14:20:55 crc kubenswrapper[4578]: I1003 14:20:55.735356 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vcftq_70df60c7-212d-4834-b237-527ca5710600/cp-frr-files/0.log" Oct 03 14:20:55 crc kubenswrapper[4578]: I1003 14:20:55.923698 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vcftq_70df60c7-212d-4834-b237-527ca5710600/cp-reloader/0.log" Oct 03 14:20:55 crc kubenswrapper[4578]: I1003 14:20:55.959442 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vcftq_70df60c7-212d-4834-b237-527ca5710600/cp-metrics/0.log" Oct 03 14:20:55 crc kubenswrapper[4578]: I1003 14:20:55.963941 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vcftq_70df60c7-212d-4834-b237-527ca5710600/cp-frr-files/0.log" Oct 03 14:20:56 crc kubenswrapper[4578]: I1003 14:20:56.019427 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vcftq_70df60c7-212d-4834-b237-527ca5710600/cp-reloader/0.log" Oct 03 14:20:56 crc kubenswrapper[4578]: I1003 14:20:56.245787 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vcftq_70df60c7-212d-4834-b237-527ca5710600/cp-metrics/0.log" Oct 03 14:20:56 crc kubenswrapper[4578]: I1003 14:20:56.279251 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vcftq_70df60c7-212d-4834-b237-527ca5710600/cp-frr-files/0.log" Oct 03 14:20:56 crc kubenswrapper[4578]: I1003 14:20:56.289094 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vcftq_70df60c7-212d-4834-b237-527ca5710600/cp-reloader/0.log" Oct 03 14:20:56 crc kubenswrapper[4578]: I1003 14:20:56.329174 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vcftq_70df60c7-212d-4834-b237-527ca5710600/cp-metrics/0.log" Oct 03 14:20:56 crc kubenswrapper[4578]: I1003 14:20:56.737951 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vcftq_70df60c7-212d-4834-b237-527ca5710600/cp-reloader/0.log" Oct 03 14:20:56 crc kubenswrapper[4578]: I1003 14:20:56.783176 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vcftq_70df60c7-212d-4834-b237-527ca5710600/cp-frr-files/0.log" Oct 03 14:20:56 crc kubenswrapper[4578]: I1003 14:20:56.838102 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vcftq_70df60c7-212d-4834-b237-527ca5710600/cp-metrics/0.log" Oct 03 14:20:56 crc kubenswrapper[4578]: I1003 14:20:56.929768 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vcftq_70df60c7-212d-4834-b237-527ca5710600/controller/0.log" Oct 03 14:20:57 crc kubenswrapper[4578]: I1003 14:20:57.123195 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vcftq_70df60c7-212d-4834-b237-527ca5710600/frr-metrics/0.log" Oct 03 14:20:57 crc kubenswrapper[4578]: I1003 14:20:57.184376 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vcftq_70df60c7-212d-4834-b237-527ca5710600/kube-rbac-proxy/0.log" Oct 03 14:20:57 crc kubenswrapper[4578]: I1003 14:20:57.353855 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vcftq_70df60c7-212d-4834-b237-527ca5710600/kube-rbac-proxy-frr/0.log" Oct 03 14:20:57 crc kubenswrapper[4578]: I1003 14:20:57.420958 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vcftq_70df60c7-212d-4834-b237-527ca5710600/reloader/0.log" Oct 03 14:20:57 crc kubenswrapper[4578]: I1003 14:20:57.707617 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-t7dz4_3d7fbaf2-ba49-481f-a56c-df68ef840423/frr-k8s-webhook-server/0.log" Oct 03 14:20:57 crc kubenswrapper[4578]: I1003 14:20:57.888314 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7969464d44-s7mk5_ce9289ad-1745-49b0-af1b-e37b45c999fa/manager/0.log" Oct 03 14:20:58 crc kubenswrapper[4578]: I1003 14:20:58.009355 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7b7dc59bdd-xqgvb_df783eb1-661a-455b-94b2-4045de8dadc0/webhook-server/0.log" Oct 03 14:20:58 crc kubenswrapper[4578]: I1003 14:20:58.233647 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vcftq_70df60c7-212d-4834-b237-527ca5710600/frr/0.log" Oct 03 14:20:58 crc kubenswrapper[4578]: I1003 14:20:58.257597 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9sc2x_c036c09c-bcdf-4fcd-908a-4f59ff7de9d9/kube-rbac-proxy/0.log" Oct 03 14:20:58 crc kubenswrapper[4578]: I1003 14:20:58.693927 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9sc2x_c036c09c-bcdf-4fcd-908a-4f59ff7de9d9/speaker/0.log" Oct 03 14:21:05 crc kubenswrapper[4578]: I1003 14:21:05.091159 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:21:05 crc kubenswrapper[4578]: I1003 14:21:05.091754 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:21:10 crc kubenswrapper[4578]: I1003 14:21:10.539731 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj_a16e5036-d08b-4217-b21b-37a28ceb701c/util/0.log" Oct 03 14:21:10 crc kubenswrapper[4578]: I1003 14:21:10.659555 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj_a16e5036-d08b-4217-b21b-37a28ceb701c/util/0.log" Oct 03 14:21:10 crc kubenswrapper[4578]: I1003 14:21:10.711879 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj_a16e5036-d08b-4217-b21b-37a28ceb701c/pull/0.log" Oct 03 14:21:10 crc kubenswrapper[4578]: I1003 14:21:10.738300 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj_a16e5036-d08b-4217-b21b-37a28ceb701c/pull/0.log" Oct 03 14:21:10 crc kubenswrapper[4578]: I1003 14:21:10.918050 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj_a16e5036-d08b-4217-b21b-37a28ceb701c/util/0.log" Oct 03 14:21:10 crc kubenswrapper[4578]: I1003 14:21:10.976556 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj_a16e5036-d08b-4217-b21b-37a28ceb701c/extract/0.log" Oct 03 14:21:11 crc kubenswrapper[4578]: I1003 14:21:11.012746 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2gzdkj_a16e5036-d08b-4217-b21b-37a28ceb701c/pull/0.log" Oct 03 14:21:11 crc kubenswrapper[4578]: I1003 14:21:11.119410 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tnrc8_70804fe7-af6b-4b8e-9731-ce7fd0f5544b/extract-utilities/0.log" Oct 03 14:21:11 crc kubenswrapper[4578]: I1003 14:21:11.339299 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tnrc8_70804fe7-af6b-4b8e-9731-ce7fd0f5544b/extract-content/0.log" Oct 03 14:21:11 crc kubenswrapper[4578]: I1003 14:21:11.340248 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tnrc8_70804fe7-af6b-4b8e-9731-ce7fd0f5544b/extract-content/0.log" Oct 03 14:21:11 crc kubenswrapper[4578]: I1003 14:21:11.378877 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tnrc8_70804fe7-af6b-4b8e-9731-ce7fd0f5544b/extract-utilities/0.log" Oct 03 14:21:11 crc kubenswrapper[4578]: I1003 14:21:11.555138 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tnrc8_70804fe7-af6b-4b8e-9731-ce7fd0f5544b/extract-utilities/0.log" Oct 03 14:21:11 crc kubenswrapper[4578]: I1003 14:21:11.580915 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tnrc8_70804fe7-af6b-4b8e-9731-ce7fd0f5544b/extract-content/0.log" Oct 03 14:21:11 crc kubenswrapper[4578]: I1003 14:21:11.845127 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qz6ml_b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39/extract-utilities/0.log" Oct 03 14:21:12 crc kubenswrapper[4578]: I1003 14:21:12.128682 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tnrc8_70804fe7-af6b-4b8e-9731-ce7fd0f5544b/registry-server/0.log" Oct 03 14:21:12 crc kubenswrapper[4578]: I1003 14:21:12.161886 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qz6ml_b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39/extract-utilities/0.log" Oct 03 14:21:12 crc kubenswrapper[4578]: I1003 14:21:12.169344 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qz6ml_b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39/extract-content/0.log" Oct 03 14:21:12 crc kubenswrapper[4578]: I1003 14:21:12.223872 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qz6ml_b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39/extract-content/0.log" Oct 03 14:21:12 crc kubenswrapper[4578]: I1003 14:21:12.387857 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qz6ml_b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39/extract-content/0.log" Oct 03 14:21:12 crc kubenswrapper[4578]: I1003 14:21:12.447765 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qz6ml_b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39/extract-utilities/0.log" Oct 03 14:21:12 crc kubenswrapper[4578]: I1003 14:21:12.728144 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx_d75a489d-f32d-4e7b-8f7a-6da21a66340b/util/0.log" Oct 03 14:21:12 crc kubenswrapper[4578]: I1003 14:21:12.972768 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx_d75a489d-f32d-4e7b-8f7a-6da21a66340b/pull/0.log" Oct 03 14:21:13 crc kubenswrapper[4578]: I1003 14:21:13.063766 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx_d75a489d-f32d-4e7b-8f7a-6da21a66340b/util/0.log" Oct 03 14:21:13 crc kubenswrapper[4578]: I1003 14:21:13.088192 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx_d75a489d-f32d-4e7b-8f7a-6da21a66340b/pull/0.log" Oct 03 14:21:13 crc kubenswrapper[4578]: I1003 14:21:13.279794 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx_d75a489d-f32d-4e7b-8f7a-6da21a66340b/util/0.log" Oct 03 14:21:13 crc kubenswrapper[4578]: I1003 14:21:13.300727 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx_d75a489d-f32d-4e7b-8f7a-6da21a66340b/pull/0.log" Oct 03 14:21:13 crc kubenswrapper[4578]: I1003 14:21:13.304341 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qz6ml_b4c28ddc-2d83-4a6a-9a3b-64ebc9949d39/registry-server/0.log" Oct 03 14:21:13 crc kubenswrapper[4578]: I1003 14:21:13.387623 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cbvznx_d75a489d-f32d-4e7b-8f7a-6da21a66340b/extract/0.log" Oct 03 14:21:13 crc kubenswrapper[4578]: I1003 14:21:13.556681 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-bf7p9_30b79cea-e428-4de1-897c-f561c1bfc659/marketplace-operator/0.log" Oct 03 14:21:13 crc kubenswrapper[4578]: I1003 14:21:13.664791 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xtdbx_90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9/extract-utilities/0.log" Oct 03 14:21:13 crc kubenswrapper[4578]: I1003 14:21:13.830956 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xtdbx_90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9/extract-content/0.log" Oct 03 14:21:13 crc kubenswrapper[4578]: I1003 14:21:13.885737 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xtdbx_90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9/extract-utilities/0.log" Oct 03 14:21:13 crc kubenswrapper[4578]: I1003 14:21:13.885737 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xtdbx_90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9/extract-content/0.log" Oct 03 14:21:14 crc kubenswrapper[4578]: I1003 14:21:14.094829 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xtdbx_90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9/extract-content/0.log" Oct 03 14:21:14 crc kubenswrapper[4578]: I1003 14:21:14.101500 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xtdbx_90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9/extract-utilities/0.log" Oct 03 14:21:14 crc kubenswrapper[4578]: I1003 14:21:14.370090 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-xtdbx_90dc2bbe-53ad-4cba-a8d2-e39bdd6519e9/registry-server/0.log" Oct 03 14:21:14 crc kubenswrapper[4578]: I1003 14:21:14.419378 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6wkpp_68798069-f345-4517-88f7-d8d50c787b5d/extract-utilities/0.log" Oct 03 14:21:14 crc kubenswrapper[4578]: I1003 14:21:14.559025 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6wkpp_68798069-f345-4517-88f7-d8d50c787b5d/extract-content/0.log" Oct 03 14:21:14 crc kubenswrapper[4578]: I1003 14:21:14.563880 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6wkpp_68798069-f345-4517-88f7-d8d50c787b5d/extract-utilities/0.log" Oct 03 14:21:14 crc kubenswrapper[4578]: I1003 14:21:14.577699 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6wkpp_68798069-f345-4517-88f7-d8d50c787b5d/extract-content/0.log" Oct 03 14:21:14 crc kubenswrapper[4578]: I1003 14:21:14.736059 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6wkpp_68798069-f345-4517-88f7-d8d50c787b5d/extract-content/0.log" Oct 03 14:21:14 crc kubenswrapper[4578]: I1003 14:21:14.855222 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6wkpp_68798069-f345-4517-88f7-d8d50c787b5d/extract-utilities/0.log" Oct 03 14:21:15 crc kubenswrapper[4578]: I1003 14:21:15.381134 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6wkpp_68798069-f345-4517-88f7-d8d50c787b5d/registry-server/0.log" Oct 03 14:21:35 crc kubenswrapper[4578]: I1003 14:21:35.091572 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:21:35 crc kubenswrapper[4578]: I1003 14:21:35.093082 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:22:05 crc kubenswrapper[4578]: I1003 14:22:05.091941 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:22:05 crc kubenswrapper[4578]: I1003 14:22:05.092481 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:22:05 crc kubenswrapper[4578]: I1003 14:22:05.092528 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 14:22:05 crc kubenswrapper[4578]: I1003 14:22:05.093259 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"99c6af77c4906d954ca12c7460c666fc7a526bfb23a532ab77b5add6bbd7dead"} pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 14:22:05 crc kubenswrapper[4578]: I1003 14:22:05.093303 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" containerID="cri-o://99c6af77c4906d954ca12c7460c666fc7a526bfb23a532ab77b5add6bbd7dead" gracePeriod=600 Oct 03 14:22:05 crc kubenswrapper[4578]: I1003 14:22:05.833979 4578 generic.go:334] "Generic (PLEG): container finished" podID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerID="99c6af77c4906d954ca12c7460c666fc7a526bfb23a532ab77b5add6bbd7dead" exitCode=0 Oct 03 14:22:05 crc kubenswrapper[4578]: I1003 14:22:05.834033 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerDied","Data":"99c6af77c4906d954ca12c7460c666fc7a526bfb23a532ab77b5add6bbd7dead"} Oct 03 14:22:05 crc kubenswrapper[4578]: I1003 14:22:05.834340 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerStarted","Data":"3581fba3d9ccb34228abf6d758890d68b21db908652079203a4276c7c62b9563"} Oct 03 14:22:05 crc kubenswrapper[4578]: I1003 14:22:05.834362 4578 scope.go:117] "RemoveContainer" containerID="fa26ce42e7cbc03aab34dd320cb53954f5b159b771723ec5ec54de99e0400577" Oct 03 14:23:25 crc kubenswrapper[4578]: I1003 14:23:25.549059 4578 generic.go:334] "Generic (PLEG): container finished" podID="c7e530e9-622a-4d8a-b2aa-88d8af716b77" containerID="0b171dd5a470b945fed5a6e01b36368ba15e66a41647104cb565ecccbadf8fa3" exitCode=0 Oct 03 14:23:25 crc kubenswrapper[4578]: I1003 14:23:25.549140 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mcgpg/must-gather-2rhkb" event={"ID":"c7e530e9-622a-4d8a-b2aa-88d8af716b77","Type":"ContainerDied","Data":"0b171dd5a470b945fed5a6e01b36368ba15e66a41647104cb565ecccbadf8fa3"} Oct 03 14:23:25 crc kubenswrapper[4578]: I1003 14:23:25.551581 4578 scope.go:117] "RemoveContainer" containerID="0b171dd5a470b945fed5a6e01b36368ba15e66a41647104cb565ecccbadf8fa3" Oct 03 14:23:26 crc kubenswrapper[4578]: I1003 14:23:26.383573 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mcgpg_must-gather-2rhkb_c7e530e9-622a-4d8a-b2aa-88d8af716b77/gather/0.log" Oct 03 14:23:34 crc kubenswrapper[4578]: I1003 14:23:34.616060 4578 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mcgpg/must-gather-2rhkb"] Oct 03 14:23:34 crc kubenswrapper[4578]: I1003 14:23:34.616863 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-mcgpg/must-gather-2rhkb" podUID="c7e530e9-622a-4d8a-b2aa-88d8af716b77" containerName="copy" containerID="cri-o://f8bcbc6e4e9f8d5b48f159903251abf4f22e825103a8665e3592c7b5992b1ca7" gracePeriod=2 Oct 03 14:23:34 crc kubenswrapper[4578]: I1003 14:23:34.627480 4578 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mcgpg/must-gather-2rhkb"] Oct 03 14:23:35 crc kubenswrapper[4578]: I1003 14:23:35.166549 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mcgpg_must-gather-2rhkb_c7e530e9-622a-4d8a-b2aa-88d8af716b77/copy/0.log" Oct 03 14:23:35 crc kubenswrapper[4578]: I1003 14:23:35.167035 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mcgpg/must-gather-2rhkb" Oct 03 14:23:35 crc kubenswrapper[4578]: I1003 14:23:35.295094 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c7e530e9-622a-4d8a-b2aa-88d8af716b77-must-gather-output\") pod \"c7e530e9-622a-4d8a-b2aa-88d8af716b77\" (UID: \"c7e530e9-622a-4d8a-b2aa-88d8af716b77\") " Oct 03 14:23:35 crc kubenswrapper[4578]: I1003 14:23:35.295266 4578 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjp46\" (UniqueName: \"kubernetes.io/projected/c7e530e9-622a-4d8a-b2aa-88d8af716b77-kube-api-access-qjp46\") pod \"c7e530e9-622a-4d8a-b2aa-88d8af716b77\" (UID: \"c7e530e9-622a-4d8a-b2aa-88d8af716b77\") " Oct 03 14:23:35 crc kubenswrapper[4578]: I1003 14:23:35.310304 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7e530e9-622a-4d8a-b2aa-88d8af716b77-kube-api-access-qjp46" (OuterVolumeSpecName: "kube-api-access-qjp46") pod "c7e530e9-622a-4d8a-b2aa-88d8af716b77" (UID: "c7e530e9-622a-4d8a-b2aa-88d8af716b77"). InnerVolumeSpecName "kube-api-access-qjp46". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 03 14:23:35 crc kubenswrapper[4578]: I1003 14:23:35.397537 4578 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjp46\" (UniqueName: \"kubernetes.io/projected/c7e530e9-622a-4d8a-b2aa-88d8af716b77-kube-api-access-qjp46\") on node \"crc\" DevicePath \"\"" Oct 03 14:23:35 crc kubenswrapper[4578]: I1003 14:23:35.503598 4578 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7e530e9-622a-4d8a-b2aa-88d8af716b77-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "c7e530e9-622a-4d8a-b2aa-88d8af716b77" (UID: "c7e530e9-622a-4d8a-b2aa-88d8af716b77"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 03 14:23:35 crc kubenswrapper[4578]: I1003 14:23:35.600993 4578 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c7e530e9-622a-4d8a-b2aa-88d8af716b77-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 03 14:23:35 crc kubenswrapper[4578]: I1003 14:23:35.629829 4578 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mcgpg_must-gather-2rhkb_c7e530e9-622a-4d8a-b2aa-88d8af716b77/copy/0.log" Oct 03 14:23:35 crc kubenswrapper[4578]: I1003 14:23:35.630181 4578 generic.go:334] "Generic (PLEG): container finished" podID="c7e530e9-622a-4d8a-b2aa-88d8af716b77" containerID="f8bcbc6e4e9f8d5b48f159903251abf4f22e825103a8665e3592c7b5992b1ca7" exitCode=143 Oct 03 14:23:35 crc kubenswrapper[4578]: I1003 14:23:35.630250 4578 scope.go:117] "RemoveContainer" containerID="f8bcbc6e4e9f8d5b48f159903251abf4f22e825103a8665e3592c7b5992b1ca7" Oct 03 14:23:35 crc kubenswrapper[4578]: I1003 14:23:35.630330 4578 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mcgpg/must-gather-2rhkb" Oct 03 14:23:35 crc kubenswrapper[4578]: I1003 14:23:35.648378 4578 scope.go:117] "RemoveContainer" containerID="0b171dd5a470b945fed5a6e01b36368ba15e66a41647104cb565ecccbadf8fa3" Oct 03 14:23:35 crc kubenswrapper[4578]: I1003 14:23:35.684848 4578 scope.go:117] "RemoveContainer" containerID="f8bcbc6e4e9f8d5b48f159903251abf4f22e825103a8665e3592c7b5992b1ca7" Oct 03 14:23:35 crc kubenswrapper[4578]: E1003 14:23:35.685339 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8bcbc6e4e9f8d5b48f159903251abf4f22e825103a8665e3592c7b5992b1ca7\": container with ID starting with f8bcbc6e4e9f8d5b48f159903251abf4f22e825103a8665e3592c7b5992b1ca7 not found: ID does not exist" containerID="f8bcbc6e4e9f8d5b48f159903251abf4f22e825103a8665e3592c7b5992b1ca7" Oct 03 14:23:35 crc kubenswrapper[4578]: I1003 14:23:35.685401 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8bcbc6e4e9f8d5b48f159903251abf4f22e825103a8665e3592c7b5992b1ca7"} err="failed to get container status \"f8bcbc6e4e9f8d5b48f159903251abf4f22e825103a8665e3592c7b5992b1ca7\": rpc error: code = NotFound desc = could not find container \"f8bcbc6e4e9f8d5b48f159903251abf4f22e825103a8665e3592c7b5992b1ca7\": container with ID starting with f8bcbc6e4e9f8d5b48f159903251abf4f22e825103a8665e3592c7b5992b1ca7 not found: ID does not exist" Oct 03 14:23:35 crc kubenswrapper[4578]: I1003 14:23:35.685434 4578 scope.go:117] "RemoveContainer" containerID="0b171dd5a470b945fed5a6e01b36368ba15e66a41647104cb565ecccbadf8fa3" Oct 03 14:23:35 crc kubenswrapper[4578]: E1003 14:23:35.686112 4578 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b171dd5a470b945fed5a6e01b36368ba15e66a41647104cb565ecccbadf8fa3\": container with ID starting with 0b171dd5a470b945fed5a6e01b36368ba15e66a41647104cb565ecccbadf8fa3 not found: ID does not exist" containerID="0b171dd5a470b945fed5a6e01b36368ba15e66a41647104cb565ecccbadf8fa3" Oct 03 14:23:35 crc kubenswrapper[4578]: I1003 14:23:35.686147 4578 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b171dd5a470b945fed5a6e01b36368ba15e66a41647104cb565ecccbadf8fa3"} err="failed to get container status \"0b171dd5a470b945fed5a6e01b36368ba15e66a41647104cb565ecccbadf8fa3\": rpc error: code = NotFound desc = could not find container \"0b171dd5a470b945fed5a6e01b36368ba15e66a41647104cb565ecccbadf8fa3\": container with ID starting with 0b171dd5a470b945fed5a6e01b36368ba15e66a41647104cb565ecccbadf8fa3 not found: ID does not exist" Oct 03 14:23:36 crc kubenswrapper[4578]: I1003 14:23:36.920549 4578 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7e530e9-622a-4d8a-b2aa-88d8af716b77" path="/var/lib/kubelet/pods/c7e530e9-622a-4d8a-b2aa-88d8af716b77/volumes" Oct 03 14:24:05 crc kubenswrapper[4578]: I1003 14:24:05.091677 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:24:05 crc kubenswrapper[4578]: I1003 14:24:05.092228 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:24:35 crc kubenswrapper[4578]: I1003 14:24:35.091725 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:24:35 crc kubenswrapper[4578]: I1003 14:24:35.092305 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:25:05 crc kubenswrapper[4578]: I1003 14:25:05.091940 4578 patch_prober.go:28] interesting pod/machine-config-daemon-qq9jm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 03 14:25:05 crc kubenswrapper[4578]: I1003 14:25:05.092561 4578 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 03 14:25:05 crc kubenswrapper[4578]: I1003 14:25:05.092617 4578 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" Oct 03 14:25:05 crc kubenswrapper[4578]: I1003 14:25:05.093428 4578 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3581fba3d9ccb34228abf6d758890d68b21db908652079203a4276c7c62b9563"} pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 03 14:25:05 crc kubenswrapper[4578]: I1003 14:25:05.093486 4578 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerName="machine-config-daemon" containerID="cri-o://3581fba3d9ccb34228abf6d758890d68b21db908652079203a4276c7c62b9563" gracePeriod=600 Oct 03 14:25:05 crc kubenswrapper[4578]: E1003 14:25:05.216674 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:25:05 crc kubenswrapper[4578]: I1003 14:25:05.465640 4578 generic.go:334] "Generic (PLEG): container finished" podID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" containerID="3581fba3d9ccb34228abf6d758890d68b21db908652079203a4276c7c62b9563" exitCode=0 Oct 03 14:25:05 crc kubenswrapper[4578]: I1003 14:25:05.465678 4578 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" event={"ID":"b7d7e663-26ba-48e2-849e-f7c9554fd7a7","Type":"ContainerDied","Data":"3581fba3d9ccb34228abf6d758890d68b21db908652079203a4276c7c62b9563"} Oct 03 14:25:05 crc kubenswrapper[4578]: I1003 14:25:05.465735 4578 scope.go:117] "RemoveContainer" containerID="99c6af77c4906d954ca12c7460c666fc7a526bfb23a532ab77b5add6bbd7dead" Oct 03 14:25:05 crc kubenswrapper[4578]: I1003 14:25:05.466559 4578 scope.go:117] "RemoveContainer" containerID="3581fba3d9ccb34228abf6d758890d68b21db908652079203a4276c7c62b9563" Oct 03 14:25:05 crc kubenswrapper[4578]: E1003 14:25:05.467142 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" Oct 03 14:25:16 crc kubenswrapper[4578]: I1003 14:25:16.909479 4578 scope.go:117] "RemoveContainer" containerID="3581fba3d9ccb34228abf6d758890d68b21db908652079203a4276c7c62b9563" Oct 03 14:25:16 crc kubenswrapper[4578]: E1003 14:25:16.911478 4578 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-qq9jm_openshift-machine-config-operator(b7d7e663-26ba-48e2-849e-f7c9554fd7a7)\"" pod="openshift-machine-config-operator/machine-config-daemon-qq9jm" podUID="b7d7e663-26ba-48e2-849e-f7c9554fd7a7" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515067756332024463 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015067756332017400 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015067742630016517 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015067742630015467 5ustar corecore